Get the latest tech news

Un Ministral, Des Ministraux


Introducing the world’s best edge models.

These models set a new frontier in knowledge, commonsense, reasoning, function-calling, and efficiency in the sub-10B category, and can be used or tuned to a variety of uses, from orchestrating agentic workflows to creating specialist task workers. Both models support up to 128k context length (currently 32k on vLLM) and Ministral 8B has a special interleaved sliding-window attention pattern for faster and memory-efficient inference. They can be tuned to handle input parsing, task routing, and calling APIs based on user intent across multiple contexts at extremely low latency and cost.

Get the Android app

Or read this on Hacker News

Read more on:

Photo of Des Ministraux

Des Ministraux

Photo of Un Ministral

Un Ministral