Create a deployment. Create a new inference deployment related to a specific model.
The request CreateDeploymentRequest
A Promise of Deployment
Create an endpoint. Create a new Endpoint related to a specific deployment.
The request CreateEndpointRequest
A Promise of Endpoint
Import a model. Import a new model to your model library.
The request CreateModelRequest
A Promise of Model
Delete a deployment. Delete an existing inference deployment.
The request DeleteDeploymentRequest
A Promise of Deployment
Delete an endpoint. Delete an existing Endpoint.
The request DeleteEndpointRequest
Delete a model. Delete an existing model from your model library.
The request DeleteModelRequest
Get a deployment. Get the deployment for the given ID.
The request GetDeploymentRequest
A Promise of Deployment
Get the CA certificate. Get the CA certificate used for the deployment of private endpoints. The CA certificate will be returned as a PEM file.
The request GetDeploymentCertificateRequest
A Promise of Blob
Get a model. Get the model for the given ID.
The request GetModelRequest
A Promise of Model
List inference deployments. List all your inference deployments.
The request ListDeploymentsRequest
A Promise of ListDeploymentsResponse
List models. List all available models.
The request ListModelsRequest
A Promise of ListModelsResponse
List available node types. List all available node types. By default, the
node types returned in the list are ordered by creation date in ascending
order, though this can be modified via the order_by field.
The request ListNodeTypesRequest
A Promise of ListNodeTypesResponse
Update a deployment. Update an existing inference deployment.
The request UpdateDeploymentRequest
A Promise of Deployment
Update an endpoint. Update an existing Endpoint.
The request UpdateEndpointRequest
A Promise of Endpoint
Waits for Deployment to be in a final state.
The request GetDeploymentRequest
Optionaloptions: Readonly<WaitForOptions<Inference.v1.Deployment>>The waiting options
A Promise of Deployment
Waits for Model to be in a final state.
The request GetModelRequest
Optionaloptions: Readonly<WaitForOptions<Inference.v1.Model>>The waiting options
A Promise of Model
Managed Inference API.
This API allows you to handle your Managed Inference services.