Gradient Deployments helps you perform effortless model serving. Easily deploy your machine learning model as an API endpoint in a few simple steps. Stop worrying about Kubernetes, Docker, and framework headaches.
Gradient Deployments is currently in Preview. If you would like to provide feedback on the deployments product, please reach out.
After Notebooks and Workflows, the third major component of the Gradient Platform is Deployments.
Deployments is used to run container images and to serve machine learning models using a high-performance, low-latency micro-service with a RESTful API.
This allows the model to be run on new unseen data in production, also known as model inference.
Where to start
The best place to start learning how to deploy models on Gradient is the official Gradient Deployments Tutorial.
Deployments has a number of concepts to allow different configuration.
A deployment can have multiple runs at the same time. Any update to the deployment spec can create a new deployment run. Once the latest deployment run is ready, the previous deployment run will scale down.