Models / Mistral AI
Code
Chat

Magistral Small 2506

24B‑parameter open‑source reasoning model from Mistral AI, fine‑tuned and RL‑trained for strong math, coding, and multilingual reasoning,

About model

Magistral Small 2506 is a small, efficient reasoning model with 24B parameters, capable of long chains of reasoning traces and supporting dozens of languages. It is suitable for local deployment and offers open usage under the Apache 2.0 License. Ideal for applications requiring multilingual reasoning capabilities.

To run this model, you first need to deploy it on a Dedicated Endpoint.

Performance benchmarks

Model

AIME 2025

GPQA Diamond

HLE

LiveCodeBench

MATH500

SWE-bench verified

48.4%

Related open-source models

Competitor closed-source models

Claude Opus 4.6

90.5%

34.2%

78.7%

OpenAI o3

83.3%

24.9%

99.2%

62.3%

OpenAI o1

76.8%

96.4%

48.9%

GPT-4o

49.2%

2.7%

32.3%

89.3%

31.0%

  • Model card

    Model Card for Magistral-Small-2506

    Building upon Mistral Small 3.1 (2503), with added reasoning capabilities, undergoing SFT from Magistral Medium traces and RL on top, it's a small, efficient reasoning model with 24B parameters.

    Magistral Small can be deployed locally, fitting within a single RTX 4090 or a 32GB RAM MacBook once quantized.

    Key Features

    • Reasoning: Capable of long chains of reasoning traces before providing an answer.
    • Multilingual: Supports dozens of languages, including English, French, German, Greek, Hindi, Indonesian, Italian, Japanese, Korean, Malay, Nepali, Polish, Portuguese, Romanian, Russian, Serbian, Spanish, Swedish, Turkish, Ukrainian, Vietnamese, Arabic, Bengali, Chinese, and Farsi.
    • Apache 2.0 License: Open license allowing usage and modification for both commercial and non-commercial purposes.
    • Context Window: A 128k context window, but performance might degrade past 40k. Hence we recommend setting the maximum model length to 40k.

    Benchmark Results

    Model AIME24 pass@1 AIME25 pass@1 GPQA DiamondLivecodebench (v5)
    Magistral Medium 73.59% 64.95% 70.83% 59.36%
    Magistral Small 70.68% 62.76% 68.18% 55.84%

    Sampling parameters

    Please make sure to use:

    • top_p: 0.95
    • temperature: 0.7
    • max_tokens: 40960

    Basic Chat Template

    We highly recommend including the default system prompt used during RL for the best results, you can edit and customise it if needed for your specific use case.

    system_prompt, user_message and assistant_response are placeholders.

    We invite you to choose, depending on your use case and requirements, between keeping reasoning traces during multi-turn interactions or keeping only the final assistant response.

Related models
  • Model provider
    Mistral AI
  • Type
    Code
    Chat
  • Main use cases
    Chat
    Reasoning
  • Deployment
    On-Demand Dedicated
    Monthly Reserved
  • Parameters
    23.6B
  • Context length
    40k
  • Input modalities
    Text
  • Output modalities
    Text