Deploy

This guide provides detailed steps to deploy your models to CX, whether you prefer the convenience of a command-line interface or the flexibility of cURL requests.

Free Deployments for Open Source models

At CX, we are committed to supporting the open-source community. We proudly offer complimentary serverless deployments for any open-source AI model. If you have a model you'd like to deploy, please contact our team, and we'll ensure it's live within 24 hours.

How to Request a Deployment:

  1. Compose an Email: Send your request to [email protected].

  2. Use the Following Format:

Subject: Open Source Deployment Request

Body:
- Model Name: [Name of the model]
- Model Link: [Link to the model repository or source]
- Model Configuration: [E.g., 7B parameters, 12 layers, etc.]

1. Command Line Interface (CLI)

The cx CLI offers a streamlined approach for deploying models to CX.

Step-by-Step Deployment

  1. Login:

  1. Push image to a private registry.

  1. Deploy the model:

Optionally, to deploy a serverless app, use the following command:

2. cURL Requests

For those seeking greater flexibility, deploying a model or initiating serverless deployments can be done using cURL requests.

Standard Deployment

Execute the following command to deploy a model:

Remember: Update the -d payload to match your desired deployment configuration and insert the token from the login step in the Authorization header.

Serverless Deployment

To deploy your model in a serverless environment:

Available GPUs:

Choose from the following GPU models

  • H100_NVLINK_80GB

  • A100_NVLINK_80GB

  • A100_NVLINK

  • A100_PCIE_40GB

  • A100_PCIE_80GB

  • A40

  • RTX_A6000

  • RTX_A5000

  • RTX_A4000

  • Tesla_V100_NVLINK

Last updated