NavigationContentFooter

Managed Inference

Dive into seamless language processing with our easy-to-use LLM endpoints. Perfect for everything from data analysis to creative tasks, all with clear pricing.

Managed Inference Quickstart

Getting Started

Quickstart

Learn how to create, connect to, and delete a Managed Inference endpoint in a few steps.

View Quickstart

Concepts

Core concepts that give you a better understanding of Scaleway Managed Inference.

View Concepts

How-tos

Check our guides about creating and managing Managed Inference endpoints.

View How-tos

Additional content

Guides to help you choose a Managed Inference endpoint, understand pricing and advanced configuration.

View additional content
Managed Inference API

Learn how to create and manage your Scaleway Managed Inference endpoints through the API.

Go to Managed Inference API

Changelog

  • Managed Inference

    Changed

    Models now support longer and better conversations

    • All models on catalog now support conversations to their full context window (e.g Mixtral-8x7b up to 32K tokens, Llama3 up to 8k tokens).
    • Llama3 70B is now available in FP8 quantization, INT8 is deprecated.
    • Llama3 8b is now available in FP8 quantization, BF16 remains default.
  • Managed Inference

    Added

    Managed Inference is available in Public Beta

    Managed Inference lets you deploy generative AI models and answer prompts from European end-consumers securely. Now available in public beta!

View the full changelog
Questions?

Visit our Help Center and find the answers to your most frequent questions.

Visit Help Center
Docs APIScaleway consoleDedibox consoleScaleway LearningScaleway.comPricingBlogCarreer
© 2023-2024 – Scaleway