Deploy View


PerceptiLabs Deploy View allows you to export and deploy your model to different targets.
The view displays the following options to select the model(s) to export/deploy:
1. Model Selection Checkbox: enable this to select the model(s) for export/deployment.
2. Search bar: allows you to filter the list of models by name.
To the right of the model selection screen are the export/deployment targets that you can click:
The following subsections describe these targets:

Export Options

The current export options include:
  • TensorFlow: exports to TensorFlow's exported model format or to TensorFlow Lite.
  • FastAPI Server: generates a TensorFlow model along with a Python server app with a simple API that you can use for inference on your model.
  • PL Package: exports a zipped package containing your PerceptiLabs model that you easily can share and load.
Selecting either of these displays a popup with some or all of the following options:
  • Save to: allows you to specify the location to which the exported model files are to be placed.
  • Optimize (available for TensorFlow model exports): provides options to compress and/or quantize your exported model(s) during export. Selecting either of these options will export to TensorFlow Lite format.

Deployment Options

Select Gradio to export and deploy your model as a Gradio app.

Using the Exported/Deployed Model

After you complete the export/deployment the model can be used for inference.
See Exporting and Deploying Models for information on how to use your exported/deployed model.