ScalewaySkip to loginSkip to main contentSkip to footer section

Accelerate your AI ambition

Leverage colossal AI models thanks to the largest GPU VRAM on the market and Blackwell innovations, including enhanced NVLink interconnectivity and energy efficiency.

Make real-time reasoning a reality

B300 Ultra GPUs enhance the accuracy for complex reasoning and long-context requests.

Achieve unmatched AI economics

FP4 precision and larger memory are key drivers to reduce the cost and energy required to serve AI models.

Push the boundaries of performance

Overcome AI bottlenecks without increasing your cost per token. The HGX B300's Blackwell architecture provides "Ultra" performance and up to 2.3 TB of memory for more precise AI reasoning.

Discover our full range of GPU Instances

Frequently asked questions

What is a GPU?

A Graphical Processing Units (GPU) is a highly specialized, massively parallel processor that has evolved from a graphics renderer into the primary engine for modern high-performance computing and artificial intelligence (AI). A GPU is the computational backbone of the modern AI datacenter, providing unmatched throughput for parallelized compute tasks. GPUs like NVIDIA's H100 and B300 are essential for accelerating the training and fine-tuning of massive models, such as LLMs, and for high-throughput inference serving.

What is a cloud GPU Instance?

A cloud GPU Instance refers to a virtual computing environment that offers access to powerful GPUs over the internet.

How quickly can I get started?

To get started, you need to create an account on the Scaleway console and and have either an Owner status or IAM permissions. Simply follow the steps indicated in the console to create your GPU Instance. Learn more here.

In which regions and Availability Zones will the B300-SXM GPU Instance be available?

The B300-SXM GPU Instance will be available in the Paris region, specifically the PAR-2 Availability Zone.

How am I billed for GPU Instances consumption?

You can opt for a “pay as you go” model (by the minute), paying only for what you consume. This approach gives you the flexibility to provision resources and delete resources when needed.

Which operating systems are available?

GPU OS 12 is a specialized OS image based on Ubuntu 24.04 (Noble), optimized for GPU-accelerated workloads. It comes pre-installed with the NVIDIA driver, Docker, and NVIDIA's Container Toolkit, providing an environment for running containerized applications. This image is designed to work efficiently with NVIDIA NGC container images, enabling rapid deployment of GPU-accelerated applications, such as machine learning and data processing workloads.

Can I use GPU Instances with Kubernetes?

Yes, you can use GPU Instances with Kapsule, our managed Kubernetes solution. Kapsule will automatically manage the setup of the GPU thanks to the NVIDIA GPU Operator. Learn more here.

Do you offer Service Level Agreements (SLA)?

Yes, we offer a 99.5% SLA on all GPU Instances. Learn more here.

What type of support is available?

Scaleway offers 4 different support plans to match your needs: Basic, Advanced, Business, and Enterprise. You can find all the information, including pricing, here.

How to launch a GPU Instance using Docker?

To access the GPU with Docker on Scaleway GPU instances, use Docker with NVIDIA Container Toolkit installed to enable GPU support in containers. Make sure to deploy your container with the --gpus flag to access the GPU resources. Learn more here.