How to deploy a model on Scaleway Managed Inference
Before you start
To complete the actions presented below, you must have:
- A Scaleway account logged into the console
- Owner status or IAM permissions allowing you to perform actions in the intended Organization
- Click the AI section of the Scaleway console, and select Managed Inference from the side menu to access the Managed Inference dashboard.
- From the drop-down menu, select the geographical region where you want to create your deployment.
- Click Deploy a model to launch the model deployment wizard.
- Provide the necessary information:
- Select the desired model and quantization to use for your deployment from the available options.
- Choose the geographical region for the deployment.
- For custom models: Choose the model quantization.
- Specify the GPU Instance type to be used with your deployment.
- Choose the number of nodes for your deployment. Note that this feature is currently in Public Beta.
- Enter a name for the deployment, and optional tags.
- Configure the network connectivity settings for the deployment:
- Attach to a Private Network for secure communication and restricted availability. Choose an existing Private Network from the drop-down list, or create a new one.
- Set up Public connectivity to access resources via the public internet. Authentication by API key is enabled by default.
- Click Deploy model to launch the deployment process. Once the model is ready, it will be listed among your deployments.
See Also
Still need help?Create a support ticket