Mistral Small 3.1: Power, Speed, and Openness in a Single Model
The AI landscape is shifting toward smaller, more efficient models—and Mistral Small 3.1 is setting the new standard. With **24 billion parameters**, a **128k-token context window**, and **multimodal capabilities** (including image recognition and OCR), this open-source model delivers both power and agility. It achieves **150 tokens per second**, supports **21+ languages**, and outperforms competitors like Gemma 3 and GPT-4 Omni Mini in benchmarks.