MistralMistral SmallJan 30, 2025
Mistral: Mistral Small 3
Mistral Small 3 is a 24B-parameter language model optimized for low-latency performance across common AI tasks. Released under the Apache 2.0 license, it features both pre-trained and instruction-tuned versions designed for efficient local deployment. The model achieves 81% accuracy on the MMLU benchmark and performs competitively with larger models like Llama 3.3 70B and Qwen 32B, while operating at three times the speed on equivalent hardware. Read the blog post about the model here.
Context Window
33K
tokens
Max Output
16K
tokens
Released
Jan 30, 2025
Arena Rank
—
Capabilities
👁Vision
🧠Reasoning
🔧Tool Calling
⚡Prompt Caching
🖥Computer Use
🎨Image Generation
Supported Parameters
Frequency Penalty
Reduce repetition
Logit Bias
Adjust token weights
Max Tokens
Output length limit
min_p
Presence Penalty
Encourage new topics
Repetition Penalty
Penalize repeated tokens
Response Format
JSON mode / structured output
Seed
Deterministic outputs
Stop Sequences
Custom stop tokens
structured_outputs
Temperature
Controls randomness
Tool Choice
Control tool usage
Tool Calling
Function calling support
Top K
Top-K sampling
Top P
Nucleus sampling
Pricing Comparison
| Router | Input / 1M | Output / 1M | Cached Input / 1M |
|---|---|---|---|
| OpenRouter | $0.05 | $0.08 | — |
| Martian | $0.05 | $0.08 | — |
| DeepInfra | $0.05 | $0.08 | — |
Model IDs
OpenRouter
mistralai/mistral-small-24b-instruct-2501Hugging Facemistralai/Mistral-Small-24B-Instruct-2501 ↗
Tags
tool-calling
Related Models
Mistral#175
Mistral: Mistral Small 3.1 24B (free)
128K ctxFree/1M in
Mistral
Mistral Small Latest
33K ctx$0.10/1M in
Mistral
Mistral: Mistral Small Creative
33K ctx$0.10/1M in
Mistral
Mistral: Mistral Small 3.2 24B
131K ctx$0.06/1M in
Mistral#48
Mistral Large 3
256K ctx$0.50/1M in
Mistral#163
Mistral Large 2407
131K ctx$2.00/1M in