ScalewaySkip to loginSkip to main contentSkip to footer section

Choose the right GPU

Finding the right GPU instance can be tricky, but we've made it easy. We have a wide range of offerings available, so you can explore all your options and compare them side-by-side.

Explore our GPU options

With thousands of machines, Scaleway has one the largest fleet of GPU Instances in Europe. We give European AI startups the tools to create products that revolutionize how we work and live, without a considerable CAPEX investment.

Train & scale

H100-SXM GPU Instance

Ideal for LLM fine-tuning and larger LLM model Inference.

From €3.12/GPU/hour

Deploy this GPU
  • NVIDIA Hopper (2022)
  • 2-8 GPUs NVIDIA H100-SXM
  • 80 GB VRAM (HBM3, 3.35 TB/s)
  • 32-128 vCPUs (Sapphire Rapids)
  • 240-960 GB RAM
  • 3.2-12,8 TB ephemeral Scratch NVMe
  • 99.5% SLA

Billed per minute

Inference

L4 GPU Instance

Ideal for image, video, and LLM Inference.

From €0.75/GPU/hour

Deploy this GPU
  • NVIDIA Ada Lovelace (2022)
  • 1-8 GPUs NVIDIA L4 (PCIe 4)
  • 24 GB VRAM (GDDR6, 300GB/s)
  • 8-64 vCPUs (AMD Zen 3)
  • 48-384 GB RAM
  • 99.5% SLA

Billed per minute

Fine-tune

H100 GPU Instance

Ideal for 7B LLM fine-tuning and Inference.

From €2.73/GPU/hour

Deploy this GPU
  • NVIDIA Hopper (2022)
  • 1 GPU NVIDIA H100 (PCIe 5)
  • 80 GB VRAM (HBM2e, 2TB/s)
  • 24 vCPUs (AMD Zen 4)
  • 240 GB RAM
  • 3 TB ephemeral Scratch NVMe
  • 99.5% SLA

Billed per minute

Inference

L40S GPU Instance

Ideal for graphics, LLM Inference, and Generative AI.

From €1.14/GPU/hour

Deploy this GPU
  • NVIDIA Ada Lovelace (2022)
  • 1-8 NVIDIA L4 (PCIe 4)
  • 48 GB VRAM (GDDR6, 864 GB/s)
  • 8-64 vCPUs (AMD Zen 3)
  • 96-768 GB RAM
  • 1.6-12.8 TB ephemeral Scratch NVMe
  • 99.5% SLA

Billed per minute

Stop the guessing game: estimate your costs

Choose your plan

*
*
*
GB
Min. 10 GB
0

0

1

2

3

4

5

Flexible IP addresses can be managed independently of any Instance. Flexible routed IPv6 addresses are free of charge; you can assign up to 5 flexible routed IPv4 addresses.

Estimated cost

Option and valuePrice
ZoneParis 1
Instance1x0€
Volume10GB0€
Flexible IPv4No0€
View all prices

GPU: instance, server, or cluster?

Find your ideal AI solution

GPU Instances are perfect for workloads such as Inference LLM and AI chatbots, with flexible configurations to suit your needs.
For large-scale AI and high-performance computing (HPC) tasks, our AI Supercomputers deliver exceptional computational power and are optimized for intensive parallel processing.
Whether you need flexibility or specialized, high-performance resources, we provide the right solution for your AI and HPC workloads.

ChooseGPU-Schema-1040px-Dark.jpg

Frequently asked questions

What is a GPU?

Graphical Processing Units (GPU) are specialized hardware originally designed for rendering graphics in video games and other 3D applications. However, their massively parallel architecture makes them ideal for various high-performance computing tasks, such as deep learning, massive machine learning, data processing, scientific simulations, and more.

What is a cloud GPU Instance ?

A cloud GPU Instance refers to a virtual computing environment that offers access to powerful GPUs over the internet.

How quickly can I get started?

To get started, you need to create an account on the Scaleway console and and have either an Owner status or IAM permissions. Simply follow the steps indicated in the console to create your GPU Instance. Learn more here.

In which regions and Availability Zones are Scaleway GPU Instances available?

Our GPU Instances are available in the fallowing regions: Paris (PAR-2, and also PAR-1 for the L4 GPU Instance) and Warsaw (WAW-2). You can check the Instance availability guide to discover where each GPU Instance is available.

How am I billed for GPU Instances consumption?

You can opt for a “pay as you go” model (by the minute), paying only for what you consume. This approach gives you the flexibility to provision resources and delete resources when needed.

Can I use GPU Instances with Kubernetes?

Yes, you can use GPU Instances with Kapsule, our managed Kubernetes solution. Kapsule will automatically manage the setup of the GPU thanks to the NVIDIA GPU Operator. Learn more here.

Which operating systems are available?

GPU OS 12 is a specialized OS image based on Ubuntu 24.04 (Noble), optimized for GPU-accelerated workloads. It comes pre-installed with the NVIDIA driver, Docker, and NVIDIA's Container Toolkit, providing an environment for running containerized applications. This image is designed to work efficiently with NVIDIA NGC container images, enabling rapid deployment of GPU-accelerated applications, such as machine learning and data processing workloads.

Do you offer Service Level Agreements?

Yes, we offer a 99.5% SLA on all GPU Instances. Learn more here.

What type of support is available?

Scaleway offers 4 different support plans to match your needs: Basic, Advanced, Business, and Enterprise. You can find all the information, including pricing, here.