Add new ACLs. Add new ACL rules for a specific deployment.
The request AddDeploymentACLRulesRequest
A Promise of AddDeploymentACLRulesResponse
Create a deployment. Create a new inference deployment related to a specific model.
The request CreateDeploymentRequest
A Promise of Deployment
Create an endpoint. Create a new Endpoint related to a specific deployment.
The request CreateEndpointRequest
A Promise of Endpoint
Delete a deployment. Delete an existing inference deployment.
The request DeleteDeploymentRequest
A Promise of Deployment
Delete an exising ACL.
The request DeleteDeploymentACLRuleRequest
Delete an endpoint. Delete an existing Endpoint.
The request DeleteEndpointRequest
Get a deployment. Get the deployment for the given ID.
The request GetDeploymentRequest
A Promise of Deployment
Get the CA certificate. Get the CA certificate used for the deployment of private endpoints. The CA certificate will be returned as a PEM file.
The request GetDeploymentCertificateRequest
A Promise of Blob
Get a model. Get the model for the given ID.
The request GetModelRequest
A Promise of Model
List your ACLs. List ACLs for a specific deployment.
The request ListDeploymentACLRulesRequest
A Promise of ListDeploymentACLRulesResponse
List inference deployments. List all your inference deployments.
The request ListDeploymentsRequest
A Promise of ListDeploymentsResponse
List models. List all available models.
The request ListModelsRequest
A Promise of ListModelsResponse
List available node types. List all available node types. By default, the
node types returned in the list are ordered by creation date in ascending
order, though this can be modified via the order_by
field.
The request ListNodeTypesRequest
A Promise of ListNodeTypesResponse
Set new ACL. Set new ACL rules for a specific deployment.
The request SetDeploymentACLRulesRequest
A Promise of SetDeploymentACLRulesResponse
Update a deployment. Update an existing inference deployment.
The request UpdateDeploymentRequest
A Promise of Deployment
Update an endpoint. Update an existing Endpoint.
The request UpdateEndpointRequest
A Promise of Endpoint
Waits for Deployment to be in a final state.
The request GetDeploymentRequest
Optional
options: Readonly<WaitForOptions<LLMInference.v1beta1.Deployment>>The waiting options
A Promise of Deployment
Managed Inference API.
This API allows you to manage your Inference services.