Accelerate your AI ambition
Leverage colossal AI models thanks to the largest GPU VRAM on the market and Blackwell innovations, including enhanced NVLink interconnectivity and energy efficiency.
Manufacture intelligence at scale.

Leverage colossal AI models thanks to the largest GPU VRAM on the market and Blackwell innovations, including enhanced NVLink interconnectivity and energy efficiency.
B300 Ultra GPUs enhance the accuracy for complex reasoning and long-context requests.
FP4 precision and larger memory are key drivers to reduce the cost and energy required to serve AI models.
Overcome AI bottlenecks without increasing your cost per token. The HGX B300's Blackwell architecture provides "Ultra" performance and up to 2.3 TB of memory for more precise AI reasoning.
Fine-tune models at a large scale for unprecedented nuance, accuracy, and reasoning in text, image, and video generation.
Run complex, real-time inference for internal or customer-facing AI assistants that require significant computational power.


H100 GPU Instances
Accelerate AI applications' development with H100-SXM GPU Instances.

Render GPU Instances
Dedicated Tesla P100s for all your Machine Learning & Artificial Intelligence needs.

L4 GPU Instance
Maximize your AI infrastructure's potential with a versatile and cost-effective GPU Instance.

L40S GPU Instance
Accelerate the next generation of AI-enabled applications with the universal L40S GPU Instance.
A Graphical Processing Units (GPU) is a highly specialized, massively parallel processor that has evolved from a graphics renderer into the primary engine for modern high-performance computing and artificial intelligence (AI). A GPU is the computational backbone of the modern AI datacenter, providing unmatched throughput for parallelized compute tasks. GPUs like NVIDIA's H100 and B300 are essential for accelerating the training and fine-tuning of massive models, such as LLMs, and for high-throughput inference serving.
A cloud GPU Instance refers to a virtual computing environment that offers access to powerful GPUs over the internet.
The B300-SXM GPU Instance will be available in the Paris region, specifically the PAR-2 Availability Zone.
You can opt for a “pay as you go” model (by the minute), paying only for what you consume. This approach gives you the flexibility to provision resources and delete resources when needed.
GPU OS 12 is a specialized OS image based on Ubuntu 24.04 (Noble), optimized for GPU-accelerated workloads. It comes pre-installed with the NVIDIA driver, Docker, and NVIDIA's Container Toolkit, providing an environment for running containerized applications. This image is designed to work efficiently with NVIDIA NGC container images, enabling rapid deployment of GPU-accelerated applications, such as machine learning and data processing workloads.
Yes, we offer a 99.5% SLA on all GPU Instances. Learn more here.
Scaleway offers 4 different support plans to match your needs: Basic, Advanced, Business, and Enterprise. You can find all the information, including pricing, here.
To access the GPU with Docker on Scaleway GPU instances, use Docker with NVIDIA Container Toolkit installed to enable GPU support in containers. Make sure to deploy your container with the --gpus flag to access the GPU resources. Learn more here.