Ministral 3 14B Instruct 2512
Frontier 14B multimodal model for high-quality assistants, analytics, and multilingual reasoning.
About model
256K
Long-horizon conversations, documents, and tool-call traces in a single pass.
14B
Strong reasoning, coding, and multilingual performance in a compact frontier-tier model.
Vision
Joint understanding of documents, screenshots, and natural language queries for rich assistant experiences.
- Multimodal Assistant: Joint text–image reasoning for chat, document analysis, dashboards, and visual Q&A
- Agentic Tool Use: Native function calling and structured outputs for long-running agents and copilots
- Multilingual & Code: Broad language coverage and strong code handling for global products and developer tools
- Long-Context Reasoning: Designed for retrieval-heavy, multi-step workflows that rely on stable 256K context
Model | AIME 2025 | GPQA Diamond | HLE | LiveCodeBench | MATH500 | SWE-bench verified |
|---|---|---|---|---|---|---|
Ministral 3 14B Instruct 2512 | 85.0% | 71.2% | 64.6% | Related open-source models | Competitor closed-source models | |
90.5% | 34.2% | 78.7% | ||||
83.3% | 24.9% | 99.2% | 62.3% | |||
76.8% | 96.4% | 48.9% | ||||
49.2% | 2.7% | 32.3% | 89.3% | 31.0% |
Model card
Architecture overview:
• Dense 14B-class language model paired with a lightweight vision encoder, exposed as a single interface for text and images.
• 256K token context window designed for extended conversations, document analysis, and long-running tool-call traces.
• Instruction-tuned head focused on stable assistant behavior, schema-following, and controllable output formatting for agents and workflows.Training and performance:
• Trained on diverse multilingual, code, and web-style data to cover a wide range of reasoning and analytic tasks.
• Instruct variant optimized for dialogue, tool use, and structured outputs rather than raw pretraining perplexity.
• Positioned to compete with much larger models on many assistant, coding, and reasoning tasks while keeping latency and cost more manageable.
Applications & use cases
Assistants and agents:
• High-end multilingual assistants for support, operations, and knowledge work that require grounded, explainable responses.
• Internal copilots that coordinate retrieval, tools, and business logic to automate complex, multi-step workflows.
• Agentic systems that plan, call tools, and synthesize results into concise recommendations or actions.Knowledge, content, and multimodal workflows:
• Long-document analysis, summarization, and synthesis across technical docs, contracts, product specs, and knowledge bases using the 256K context window.
• Multimodal understanding of screenshots, diagrams, and document snippets for debugging, troubleshooting, and guided workflows.
• High-quality content drafting, editing, and transformation across many languages, including structured reports, specifications, and templated outputs.
- TypeChat
- Main use casesChatMedium General Purpose
- DeploymentOn-Demand DedicatedMonthly Reserved
- Parameters13.9B
- Context length256K
- Input price
$0.20 / 1M tokens
- Output price
$0.20 / 1M tokens
- Input modalitiesText
- Output modalitiesText
- ReleasedOctober 31, 2025
- Last updatedFebruary 24, 2026
- Quantization levelFP8
- External link
- CategoryChat