Mistral Small 3
24B model rivaling GPT-4o mini, and larger models like Llama 3.3 70B. Ideal for chat use cases like customer support, translation and summarization.
About model
Mistral Small 3 is a 24B parameter large language model with state-of-the-art capabilities, suitable for fast response conversational agents, low latency function calling, and subject matter experts via fine-tuning. It is exceptionally knowledge-dense and supports dozens of languages. Ideal for hobbyists, organizations, and enterprises handling sensitive data.
Model | AIME 2025 | GPQA Diamond | HLE | LiveCodeBench | MATH500 | SWE-bench verified |
|---|---|---|---|---|---|---|
Mistral Small 3 | 45.3% | Related open-source models | Competitor closed-source models | |||
90.5% | 34.2% | 78.7% | ||||
83.3% | 24.9% | 99.2% | 62.3% | |||
76.8% | 96.4% | 48.9% | ||||
49.2% | 2.7% | 32.3% | 89.3% | 31.0% |
API usage
Endpoint:
- TypeChat
- Main use casesChatMedium General PurposeFunction Calling
- FeaturesFunction Calling
- DeploymentServerlessOn-Demand DedicatedMonthly Reserved
- Parameters24B
- Context length32K
- Input price
$0.10 / 1M tokens
- Output price
$0.30 / 1M tokens
- Input modalitiesText
- Output modalitiesText
- ReleasedJanuary 28, 2025
- Quantization levelFP16
- External link
- CategoryChat