Skip to main content
Version: 3.2

Model Deployment

Model deployment is the process of putting machine learning models into production. This makes the model’s predictions available to users, developers or systems, so they can make business decisions based on data, interact with their application (like recognize a face in an image) and so on. If you want your model to serve another application, you will want to serve it in the form of an API endpoint. Katonic Model APIs are scalable REST APIs that can create an endpoint from any function in Python. The Katonic Model APIs are commonly used when you need an API to query your model in near real-time. Model APIs are built as docker images and deployed on Katonic. You can export the model images to your external container registry and deploy them in any other hosting environment outside of Katonic using your custom CI/CD pipeline. Katonic supports REST APIs that enable you to programmatically build new model images on Katonic and export them to your external container registry.

To deploy your model with Katonic Model API deployment

1.1. Navigate to deploy section from sidebar on the platform.

Untitled

1.2. Click on ‘Create Deployment’.

Untitled

1.3. Select ‘Model API’ from the dropdown.

Untitled

1.4. Fill in your model details in the dialog box.

  • Give a name to your deployment and proceed to the next field.

Untitled

  • Select a model from the dropdown list.

Untitled

  • Select a model version from the dropdown.

Untitled

  • Select the appropriate model type.

Untitled

  • Select cluster size of the resources required for your model deployment from the dropdown.

Untitled

  • Define the maximum and minimum number of pods required for deployment and proceed to click on Deploy

Untitled

  • Once your Model API is deployed you will be able to view it in the Deploy section where it will be in "Processing" state. Click on refresh to update the status.

Untitled

  • Once your Model API is in "Running" state you can access the API Token by clicking on API.

Untitled

  • Set the expiry time and generate your API Token from the pop-up dialog box

Untitled

  • Copy the generated API Token.

Untitled

  • You can also modify the resources for your Model API by clicking ‘Update’ on the deployment panel and updating the resources to your requirement.

Untitled

  • Click on ‘Monitor’ to monitor the effectiveness and efficiency of your deployed model.

Untitled

  • In ‘Katonic Model Monitoring’ that opens up in a new tab, you can get real-time insights and alerts on model performance and data characteristics. You can also debug anomalies and initiate trigger to execute ML production pipelines to retrain the models with new data, depending on your use case.

Untitled

  • To delete an existing model deployment, click on the bin icon in your deployment panel and proceed to click on delete.

Untitled