MistralMixtralOpen SourceArena #221Apr 17, 2024
Mistral: Mixtral 8x22B Instruct
Mistral's official instruct fine-tuned version of Mixtral 8x22B. It uses 39B active parameters out of 141B, offering unparalleled cost efficiency for its size. Its strengths include: - strong math, coding, and reasoning - large context length (64k) - fluency in English, French, Italian, German, and Spanish See benchmarks on the launch announcement here. #moe
Context Window
66K
tokens
Max Output
—
tokens
Released
Apr 17, 2024
Arena Rank
#221
of 305 models
Capabilities
👁Vision
🧠Reasoning
🔧Tool Calling
⚡Prompt Caching
🖥Computer Use
🎨Image Generation
Supported Parameters
Frequency Penalty
Reduce repetition
Max Tokens
Output length limit
Presence Penalty
Encourage new topics
Response Format
JSON mode / structured output
Seed
Deterministic outputs
Stop Sequences
Custom stop tokens
structured_outputs
Temperature
Controls randomness
Tool Choice
Control tool usage
Tool Calling
Function calling support
Top P
Nucleus sampling
Pricing Comparison
| Router | Input / 1M | Output / 1M | Cached Input / 1M |
|---|---|---|---|
| OpenRouter | $2.00 | $6.00 | — |
| Vercel AI | $1.20 | $1.20 | — |
| Martian | $2.00 | $6.00 | — |
Benchmarks
Open LLM Leaderboard
AverageOpen LLM Leaderboard
33.89/100IFEvalOpen LLM Leaderboard
71.84/100BBHOpen LLM Leaderboard
44.11/100MATH Lvl 5Open LLM Leaderboard
18.73/100GPQAOpen LLM Leaderboard
16.44/100MUSROpen LLM Leaderboard
13.49/100MMLU-PROOpen LLM Leaderboard
38.7/100Model IDs
OpenRouter
mistralai/mixtral-8x22b-instructHugging Facemistralai/Mixtral-8x22B-Instruct-v0.1 ↗
Tags
tool-calling
Related Models
Mistral#238
Mistral: Mixtral 8x7B Instruct
33K ctx$0.54/1M in
Mistral#48
Mistral Large 3
256K ctx$0.50/1M in
Mistral#163
Mistral Large 2407
131K ctx$2.00/1M in
Mistral#173
Mistral Large 2411
131K ctx$2.00/1M in
Mistral#174
Magistral Medium 2509
128K ctx$2.00/1M in
Mistral#175
Mistral: Mistral Small 3.1 24B (free)
128K ctxFree/1M in