Inference API
- POST🆕 Inference by environment
- POSTProduction deployment
- POSTDevelopment deployment
- POSTPublished deployment
- OpenAI compatible endpoints
- Wake deployment endpoints
- Deprecated endpoints
Async Inference API
Management API
- GETGet all secrets
- POSTUpsert a secret
- 🆕 Manage model environments
- 🆕 Manage chain environments
- GETGet instance types
- Get models
- DELDelete models
- Get chains
- DELDelete chains
- Get model deployments
- DELDelete model deployments
- Get chain deployments
- DELDelete chain deployments
- Promote deployment
- Update model deployment autoscaling settings
- Activate model deployment
- Deactivate model deployment
Get all model deployments
Authorizations
You must specify the scheme 'Api-Key' in the Authorization header. For example, Authorization: Api-Key <Your_Api_Key>
Path Parameters
Response
A list of deployments of a model.
A list of deployments of a model
Unique identifier of the deployment
Time the deployment was created in ISO 8601 format
Name of the deployment
Unique identifier of the model
Whether the deployment is the production deployment of the model
Whether the deployment is the development deployment of the model
Status of the deployment
BUILDING
, DEPLOYING
, DEPLOY_FAILED
, LOADING_MODEL
, ACTIVE
, UNHEALTHY
, BUILD_FAILED
, BUILD_STOPPED
, DEACTIVATING
, INACTIVE
, FAILED
, UPDATING
, SCALED_TO_ZERO
, WAKING_UP
Number of active replicas
Autoscaling settings for the deployment. If null, the model has not finished deploying
Minimum number of replicas
Maximum number of replicas
Timeframe of traffic considered for autoscaling decisions
Waiting period before scaling down any active replica
Number of requests per replica before scaling up
The environment associated with the deployment