Models / Chat / Mistral Small 3 API
Mistral Small 3 API
Chat
24B model rivaling GPT-4o mini, and larger models like Llama 3.3 70B. Ideal for chat use cases like customer support, translation and summarization.
Try our Mistral Small 3 API

API Usage
How to use Mistral Small 3Model CardPrompting Mistral Small 3Applications & Use CasesHow to use Mistral Small 3Mistral Small 3 API Usage
Endpoint
mistralai/Mistral-Small-24B-Instruct-2501
RUN INFERENCE
import Together from "together-ai";
const together = new Together();
const response = await together.chat.completions.create({
messages: [
{
role: "user",
content: "What are some fun things to do in New York?"
}
],
model: "meta-llama/Llama-Vision-Free"
});
console.log(response.choices[0].message.content)
JSON RESPONSE
RUN INFERENCE
from together import Together
client = Together()
response = client.chat.completions.create(
model="mistralai/Mistral-Small-24B-Instruct-2501",
messages=[
{
"role": "user",
"content": "What are some fun things to do in New York?"
}
]
)
print(response.choices[0].message.content)
JSON RESPONSE
RUN INFERENCE
import Together from "together-ai";
const together = new Together();
const response = await together.chat.completions.create({
messages: [
{
role: "user",
content: "What are some fun things to do in New York?"
}
],
model: "mistralai/Mistral-Small-24B-Instruct-2501"
});
console.log(response.choices[0].message.content)
JSON RESPONSE
Model Provider:
Mistral AI
Type:
Chat
Variant:
Small
Parameters:
24B
Deployment:
✔ Serverless
✔️ On-Demand Dedicated
Quantization
FP16
Context length:
32,768
Pricing:
$0.80
Check pricing
Run in playground
Deploy model
Quickstart docs
Quickstart docs
How to use Mistral Small 3
Model details
Prompting Mistral Small 3
Applications & Use Cases
How to use Mistral Small 3
Looking for production scale? Deploy on a dedicated endpoint
Deploy Mistral Small 3 on a dedicated endpoint with custom hardware configuration, as many instances as you need, and auto-scaling.
