How to import custom models into Managed Inference
Scaleway provides a selection of common models for deployment from the Scaleway console. If you need a specific model, you can import it into Managed Inference directly from Hugging Face or a Scaleway Object Storage bucket.
Before you start
To complete the actions presented below, you must have:
- A Scaleway account logged into the console.
- Owner status or IAM permissions to perform actions in your Organization.
- Click Managed Inference in the AI section of the side menu in the Scaleway console to access the dashboard.
- From the drop-down menu, select the geographical region you want to manage.
- Click Deploy a model to launch the model deployment wizard.
- In the Choose a model section, select Custom model. If you have no model yet, click Import a model to start the model import wizard.
- Choose an upload source:
- Hugging Face: Pull the model from Hugging Face.
- Object Storage: This feature is coming soon.
- Enter your Hugging Face access token, which must have READ access to the repository.
- Enter the name of the Hugging Face repository to pull the model from.
- Choose a name for your model. The name must be unique within your Organization and Project and cannot be changed later.
- Click Verify import to check your Hugging Face credentials and ensure model compatibility.
- Review the summary of your import, which includes:
- Context size by node type.
- Quantization options.
- Estimated cost. Once checked, click Begin import to finalize the process.
Your imported model will now appear in the model library and you can deploy it on Managed Inference.
See Also
Still need help?Create a support ticket