Choose among ready-to-be-served LLMs
What makes inference fast? Model optimization is one lever. To be served fast, a model must be optimized to the machines that run it.
This isn't always a piece of cake, and can turn into a time-consuming process. That's why Scaleway is providing an evolutionary Model Library, with curated and optimized LLMs.