Mistral AI recently unveiled Magistral‑Small‑2506, a 24‑billion‑parameter reasoning model built on Mistral‑Small‑3.1‑2503, enhanced via supervised fine‑tuning and reinforcement learning traces from its larger sibling, Magistral Medium (huggingface.co). Designed with a focus on clarity, step‑by‑step deduction, and multilingual support, it offers reasoning capabilities on par with larger models, yet remains remarkably efficient.
Magistral‑Small holds its own against larger models:
| Benchmark | Magistral Medium | Magistral Small |
|---|---|---|
| AIME‑24 (pass@1) | 73.6% | 70.7% |
| AIME‑25 | 64.9% | 62.8% |
| LiveCodeBench v5 | 59.4% | 55.8% |
| GPQA Diamond | 70.8% | 68.2% |
While it trails slightly behind Medium, Magistral‑Small delivers solid performance in math, code, and STEM tasks at a fraction of the footprint (huggingface.co, mistral.ai).
Minor setup, powerful results:
temperature = 0.7, top_p = 0.95max_tokens = 40960 (40K) (1stdibs.com, huggingface.co).Under the hood, Magistral leverages a novel Reinforcement Learning from Verifiable Rewards (RLVR) stack with Mistral’s own GRPO algorithm. This approach boosts reasoning ability by ~50% on key benchmarks compared to base models, without relying on external distillation (mistral.ai). It’s a research-forward blueprint for ethical, interpretable LLM development.
mistralai/Magistral‑Small‑2506 (and GGUF version) (mistral.ai, huggingface.co).--jinja, temp 0.7, and top_p 0.95, plus 8K+ context (reddit.com).Magistral‑Small‑2506 offers a rare combination—compact, open-source reasoning excellence rivaling much larger models, powered by transparent, step-by-step logic. It’s particularly compelling for developers and researchers seeking trustworthiness, locality, and high performance in a single package.