LLM Router
HomeRoutersModelsProvidersBenchmarksPricingCompareBlogAbout
HomeModelsBenchmarksPricingCompareBlog
LLM Router

Independent comparison platform for LLM routing infrastructure.

Platform

  • Home
  • Routers
  • Models
  • Pricing
  • Blog
  • About

Routers

  • Requesty
  • OpenRouter
  • Martian
  • Unify
  • LiteLLM

© 2026 LLM Router

Data from public sources. May not reflect real-time pricing.

Providers›Mistral

Mistral

↗ Website

Mistral AI is a French AI company building open and commercial language models. Known for efficient architectures like Mixtral (mixture of experts) and strong multilingual capabilities, Mistral offers both open-source and premium API models.

Pricing available from Requesty, Vercel AI, OpenRouter, Martian, DeepInfra.

Total Models
49
Arena Ranked
11
of 49
Open Source
7
of 49
Cheapest Input
$0.02
per 1M tokens

$ Pricing Summary(per 1M tokens)

MetricInputOutput
Cheapest$0.02$0.04
Average$0.53$1.55
Most Expensive$2.00$6.00

⚙ Capabilities

👁
Vision
9
of 49 models
🧠
Reasoning
3
of 49 models
🔧
Tool Calling
37
of 49 models
⚡
Prompt Caching
3
of 49 models
🖥
Computer Use
0
of 49 models
🎨
Image Generation
0
of 49 models

🤖 All Mistral Models(49)

MistralMistral Large
#48

Mistral Large 3

Mistral Large 3 2512 is Mistral’s most capable model to date. It has a sparse mixture-of-experts architecture with 41B active parameters (675B total).

Context
256K
Max Output
256K
Input/1M
$0.50
Pricing (per 1M tokens)
Vercel AI$0.50 / $1.50
2025-12-02View details →
MistralMistral Large
#163

Mistral Large 2407

This is Mistral AI's flagship model, Mistral Large 2 (version mistral-large-2407). It's a proprietary weights-available model and excels at reasoning, code, JSON, chat, and more. Read the launch announcement [here](https://mistral.ai/news/mistral-large-2407/). It supports dozens of languages including French, German, Spanish, Italian, Portuguese, Arabic, Hindi, Russian, Chinese, Japanese, and Korean, along with 80+ coding languages including Python, Java, C, C++, JavaScript, and Bash. Its long context window allows precise information recall from large documents.

Context
131K
Max Output
—
Input/1M
$2.00
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$2.00 / $6.00
Martian$2.00 / $6.00
2024-11-19View details →
MistralMistral Large
#173

Mistral Large 2411

Mistral Large 2 2411 is an update of [Mistral Large 2](/mistralai/mistral-large) released together with [Pixtral Large 2411](/mistralai/pixtral-large-2411) It provides a significant upgrade on the previous [Mistral Large 24.07](/mistralai/mistral-large-2407), with notable improvements in long context understanding, a new system prompt, and more accurate function calling.

Context
131K
Max Output
—
Input/1M
$2.00
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$2.00 / $6.00
Martian$2.00 / $6.00
2024-11-19View details →
MistralMistral Medium
#174

Magistral Medium 2509

Complex thinking, backed by deep understanding, with transparent reasoning you can follow and verify. The model excels in maintaining high-fidelity reasoning across numerous languages, even when switching between languages mid-task.

Context
128K
Max Output
64K
Input/1M
$2.00
🧠 Reasoning
Pricing (per 1M tokens)
Vercel AI$2.00 / $5.00
2025-03-17View details →
MistralMistral Small
#175

Mistral: Mistral Small 3.1 24B (free)

Mistral Small 3.1 24B Instruct is an upgraded variant of Mistral Small 3 (2501), featuring 24 billion parameters with advanced multimodal capabilities. It provides state-of-the-art performance in text-based reasoning and vision tasks, including image analysis, programming, mathematical reasoning, and multilingual support across dozens of languages. Equipped with an extensive 128k token context window and optimized for efficient local inference, it supports use cases such as conversational agents, function calling, long-document comprehension, and privacy-sensitive deployments. The updated version is [Mistral Small 3.2](mistralai/mistral-small-3.2-24b-instruct)

Context
128K
Max Output
—
Input/1M
Free
👁 Vision🔧 Tools
Pricing (per 1M tokens)
OpenRouterFree / Free
Martian$0.03 / $0.11
2025-03-17View details →
Mistral
#215

Ministral 14B

Ministral 3 14B is the largest model in the Ministral 3 family, offering state-of-the-art capabilities and performance comparable to its larger Mistral Small 3.2 24B counterpart. Optimized for local deployment, it delivers high performance across diverse hardware, including local setups.

Context
256K
Max Output
256K
Input/1M
$0.20
Pricing (per 1M tokens)
Vercel AI$0.20 / $0.20
2025-12-01View details →
Mistral
#215

Ministral 8B

A more powerful model with faster, memory-efficient inference, ideal for complex workflows and demanding edge applications.

Context
128K
Max Output
4K
Input/1M
$0.10
🔧 Tools
Pricing (per 1M tokens)
Vercel AI$0.10 / $0.10
2024-10-01View details →
MistralMixtralOSS
#221

Mistral: Mixtral 8x22B Instruct

Mistral's official instruct fine-tuned version of [Mixtral 8x22B](/models/mistralai/mixtral-8x22b). It uses 39B active parameters out of 141B, offering unparalleled cost efficiency for its size. Its strengths include: - strong math, coding, and reasoning - large context length (64k) - fluency in English, French, Italian, German, and Spanish See benchmarks on the launch announcement [here](https://mistral.ai/news/mixtral-8x22b/). #moe

Context
66K
Max Output
—
Input/1M
$1.20
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$2.00 / $6.00
Vercel AI$1.20 / $1.20
Martian$2.00 / $6.00
2024-04-17View details →
MistralMistral Medium
#226

Mistral Medium Latest

 

Context
131K
Max Output
—
Input/1M
$0.40
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.40 / $2.00
Vercel AI$0.40 / $2.00
2025-05-07View details →
MistralMixtralOSS
#238

Mistral: Mixtral 8x7B Instruct

Mixtral 8x7B Instruct is a pretrained generative Sparse Mixture of Experts, by Mistral AI, for chat and instruction use. Incorporates 8 experts (feed-forward networks) for a total of 47 billion parameters. Instruct model fine-tuned by Mistral. #moe

Context
33K
Max Output
16K
Input/1M
$0.54
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.54 / $0.54
Martian$0.54 / $0.54
2023-12-10View details →
MistralOSS
#288

Mistral: Mistral 7B Instruct

A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length. *Mistral 7B Instruct has multiple version variants, and this is intended to be the latest version.*

Context
33K
Max Output
4K
Input/1M
$0.20
Pricing (per 1M tokens)
OpenRouter$0.20 / $0.20
Martian$0.20 / $0.20
2024-05-27View details →
MistralMistral Small

Mistral: Mistral Small Creative

Mistral Small Creative is an experimental small model designed for creative writing, narrative generation, roleplay and character-driven dialogue, general-purpose instruction following, and conversational agents.

Context
33K
Max Output
—
Input/1M
$0.10
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.10 / $0.30
Martian$0.10 / $0.30
2025-12-16View details →
Mistral

Mistral: Devstral 2 2512

Devstral 2 is a state-of-the-art open-source model by Mistral AI specializing in agentic coding. It is a 123B-parameter dense transformer model supporting a 256K context window. Devstral 2 supports exploring codebases and orchestrating changes across multiple files while maintaining architecture-level context. It tracks framework dependencies, detects failures, and retries with corrections—solving challenges like bug fixing and modernizing legacy systems. The model can be fine-tuned to prioritize specific languages or optimize for large enterprise codebases. It is available under a modified MIT license.

Context
262K
Max Output
66K
Input/1M
$0.05
🔧 Tools⚡ Cache
Pricing (per 1M tokens)
OpenRouter$0.05 / $0.22
2025-12-09View details →
Mistral

Devstral 2

An enterprise-grade text model that excels at using tools to explore codebases, editing multiple files, and powering software engineering agents.

Context
256K
Max Output
256K
Input/1M
Free
🔧 Tools
Pricing (per 1M tokens)
Vercel AIFree / Free
2025-12-09View details →
Mistral

Mistral: Ministral 3 14B 2512

The largest model in the Ministral 3 family, Ministral 3 14B offers frontier capabilities and performance comparable to its larger Mistral Small 3.2 24B counterpart. A powerful and efficient language model with vision capabilities.

Context
262K
Max Output
—
Input/1M
$0.20
👁 Vision🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.20 / $0.20
Martian$0.20 / $0.20
2025-12-02View details →
Mistral

Mistral: Ministral 3 8B 2512

A balanced model in the Ministral 3 family, Ministral 3 8B is a powerful, efficient tiny language model with vision capabilities.

Context
262K
Max Output
—
Input/1M
$0.15
👁 Vision🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.15 / $0.15
Martian$0.15 / $0.15
2025-12-02View details →
Mistral

Mistral: Ministral 3 3B 2512

The smallest model in the Ministral 3 family, Ministral 3 3B is a powerful, efficient tiny language model with vision capabilities.

Context
131K
Max Output
—
Input/1M
$0.10
👁 Vision🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.10 / $0.10
Martian$0.10 / $0.10
2025-12-02View details →
MistralMistral Large

Mistral: Mistral Large 3 2512

Mistral Large 3 2512 is Mistral’s most capable model to date, featuring a sparse mixture-of-experts architecture with 41B active parameters (675B total), and released under the Apache 2.0 license.

Context
262K
Max Output
—
Input/1M
$0.50
👁 Vision🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.50 / $1.50
Martian$0.50 / $1.50
2025-12-01View details →
Mistral

Mistral: Voxtral Small 24B 2507

Voxtral Small is an enhancement of Mistral Small 3, incorporating state-of-the-art audio input capabilities while retaining best-in-class text performance. It excels at speech transcription, translation and audio understanding. Input audio is priced at $100 per million seconds.

Context
32K
Max Output
—
Input/1M
$0.10
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.10 / $0.30
Martian$0.10 / $0.30
2025-10-30View details →
MistralMistral Medium

Mistral: Mistral Medium 3.1

Mistral Medium 3.1 is an updated version of Mistral Medium 3, which is a high-performance enterprise-grade language model designed to deliver frontier-level capabilities at significantly reduced operational cost. It balances state-of-the-art reasoning and multimodal performance with 8× lower cost compared to traditional large models, making it suitable for scalable deployments across professional and industrial use cases. The model excels in domains such as coding, STEM reasoning, and enterprise adaptation. It supports hybrid, on-prem, and in-VPC deployments and is optimized for integration into custom workflows. Mistral Medium 3.1 offers competitive accuracy relative to larger models like Claude Sonnet 3.5/3.7, Llama 4 Maverick, and Command R+, while maintaining broad compatibility across cloud environments.

Context
131K
Max Output
—
Input/1M
$0.40
👁 Vision🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.40 / $2.00
Martian$0.40 / $2.00
2025-08-13View details →
Mistral

Mistral: Codestral 2508

Mistral's cutting-edge language model for coding released end of July 2025. Codestral specializes in low-latency, high-frequency tasks such as fill-in-the-middle (FIM), code correction and test generation. [Blog Post](https://mistral.ai/news/codestral-25-08)

Context
256K
Max Output
—
Input/1M
$0.30
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.30 / $0.90
Martian$0.30 / $0.90
2025-08-01View details →
Mistral

Devstral Small Latest

Devstral Small 1.1 is a 24B parameter open-weight language model for software engineering agents, developed by Mistral AI in collaboration with All Hands AI. Finetuned from Mistral Small 3.1 and released under the Apache 2.0 license, it features a 128k token context window and supports both Mistral-style function calling and XML output formats. Designed for agentic coding workflows, Devstral Small 1.1 is optimized for tasks such as codebase exploration, multi-file edits, and integration into autonomous development agents like OpenHands and Cline. It achieves 53.6% on SWE-Bench Verified, surpassing all other open models on this benchmark, while remaining lightweight enough to run on a single 4090 GPU or Apple silicon machine. The model uses a Tekken tokenizer with a 131k vocabulary and is deployable via vLLM, Transformers, Ollama, LM Studio, and other OpenAI-compatible runtimes.

Context
131K
Max Output
—
Input/1M
$0.10
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.10 / $0.30
OpenRouter$0.10 / $0.30
Vercel AI$0.10 / $0.30
Martian$0.10 / $0.30
2025-07-10View details →
Mistral

Mistral: Devstral Medium

Devstral Medium is a high-performance code generation and agentic reasoning model developed jointly by Mistral AI and All Hands AI. Positioned as a step up from Devstral Small, it achieves 61.6% on SWE-Bench Verified, placing it ahead of Gemini 2.5 Pro and GPT-4.1 in code-related tasks, at a fraction of the cost. It is designed for generalization across prompt styles and tool use in code agents and frameworks. Devstral Medium is available via API only (not open-weight), and supports enterprise deployment on private infrastructure, with optional fine-tuning capabilities.

Context
131K
Max Output
—
Input/1M
$0.40
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.40 / $2.00
Martian$0.40 / $2.00
2025-07-10View details →
Mistral

Venice: Uncensored (free)

Venice Uncensored Dolphin Mistral 24B Venice Edition is a fine-tuned variant of Mistral-Small-24B-Instruct-2501, developed by dphn.ai in collaboration with Venice.ai. This model is designed as an “uncensored” instruct-tuned LLM, preserving user control over alignment, system prompts, and behavior. Intended for advanced and unrestricted use cases, Venice Uncensored emphasizes steerability and transparent behavior, removing default safety and alignment layers typically found in mainstream assistant models.

Context
33K
Max Output
—
Input/1M
Free
Pricing (per 1M tokens)
OpenRouterFree / Free
2025-07-09View details →
MistralMistral Small

Mistral: Mistral Small 3.2 24B

Mistral-Small-3.2-24B-Instruct-2506 is an updated 24B parameter model from Mistral optimized for instruction following, repetition reduction, and improved function calling. Compared to the 3.1 release, version 3.2 significantly improves accuracy on WildBench and Arena Hard, reduces infinite generations, and delivers gains in tool use and structured output tasks. It supports image and text inputs with structured outputs, function/tool calling, and strong performance across coding (HumanEval+, MBPP), STEM (MMLU, MATH, GPQA), and vision benchmarks (ChartQA, DocVQA).

Context
131K
Max Output
131K
Input/1M
$0.06
👁 Vision🔧 Tools⚡ Cache
Pricing (per 1M tokens)
OpenRouter$0.06 / $0.18
Martian$0.06 / $0.18
2025-06-20View details →
Mistral

Codestral Embed

Code embedding model that can embed code databases and repositories to power coding assistants.

Context
0
Max Output
—
Input/1M
$0.15
Pricing (per 1M tokens)
Vercel AI$0.15 / Free
2025-05-28View details →
Mistral

Nous: DeepHermes 3 Mistral 24B Preview

DeepHermes 3 (Mistral 24B Preview) is an instruction-tuned language model by Nous Research based on Mistral-Small-24B, designed for chat, function calling, and advanced multi-turn reasoning. It introduces a dual-mode system that toggles between intuitive chat responses and structured “deep reasoning” mode using special system prompts. Fine-tuned via distillation from R1, it supports structured output (JSON mode) and function call syntax for agent-based applications. DeepHermes 3 supports a **reasoning toggle via system prompt**, allowing users to switch between fast, intuitive responses and deliberate, multi-step reasoning. When activated with the following specific system instruction, the model enters a *"deep thinking"* mode—generating extended chains of thought wrapped in `<think></think>` tags before delivering a final answer. System Prompt: You are a deep thinking AI, you may use extremely long chains of thought to deeply consider the problem and deliberate with yourself via systematic reasoning processes to help come to a correct solution prior to answering. You should enclose your thoughts and internal monologue inside <think> </think> tags, and then provide your solution or response to the problem.

Context
33K
Max Output
33K
Input/1M
$0.02
🧠 Reasoning🔧 Tools⚡ Cache
Pricing (per 1M tokens)
OpenRouter$0.02 / $0.10
2025-05-09View details →
MistralMistral Medium

Mistral: Mistral Medium 3

Mistral Medium 3 is a high-performance enterprise-grade language model designed to deliver frontier-level capabilities at significantly reduced operational cost. It balances state-of-the-art reasoning and multimodal performance with 8× lower cost compared to traditional large models, making it suitable for scalable deployments across professional and industrial use cases. The model excels in domains such as coding, STEM reasoning, and enterprise adaptation. It supports hybrid, on-prem, and in-VPC deployments and is optimized for integration into custom workflows. Mistral Medium 3 offers competitive accuracy relative to larger models like Claude Sonnet 3.5/3.7, Llama 4 Maverick, and Command R+, while maintaining broad compatibility across cloud environments.

Context
131K
Max Output
—
Input/1M
$0.40
👁 Vision🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.40 / $2.00
Martian$0.40 / $2.00
2025-05-07View details →
Mistral

Devstral Small 2

Our open source model that excels at using tools to explore codebases, editing multiple files, and powering software engineering agents.

Context
256K
Max Output
256K
Input/1M
Free
🔧 Tools
Pricing (per 1M tokens)
Vercel AIFree / Free
2025-05-07View details →
MistralMistral Medium

Magistral Small 2509

Complex thinking, backed by deep understanding, with transparent reasoning you can follow and verify. The model excels in maintaining high-fidelity reasoning across numerous languages, even when switching between languages mid-task.

Context
128K
Max Output
64K
Input/1M
$0.50
🧠 Reasoning
Pricing (per 1M tokens)
Vercel AI$0.50 / $1.50
2025-03-17View details →
Mistral

Mistral: Saba

Mistral Saba is a 24B-parameter language model specifically designed for the Middle East and South Asia, delivering accurate and contextually relevant responses while maintaining efficient performance. Trained on curated regional datasets, it supports multiple Indian-origin languages—including Tamil and Malayalam—alongside Arabic. This makes it a versatile option for a range of regional and multilingual applications. Read more at the blog post [here](https://mistral.ai/en/news/mistral-saba)

Context
33K
Max Output
—
Input/1M
$0.20
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.20 / $0.60
Martian$0.20 / $0.60
2025-02-17View details →
MistralMistral Small

Mistral: Mistral Small 3

Mistral Small 3 is a 24B-parameter language model optimized for low-latency performance across common AI tasks. Released under the Apache 2.0 license, it features both pre-trained and instruction-tuned versions designed for efficient local deployment. The model achieves 81% accuracy on the MMLU benchmark and performs competitively with larger models like Llama 3.3 70B and Qwen 32B, while operating at three times the speed on equivalent hardware. [Read the blog post about the model here.](https://mistral.ai/news/mistral-small-3/)

Context
33K
Max Output
16K
Input/1M
$0.05
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.05 / $0.08
Martian$0.05 / $0.08
DeepInfra$0.05 / $0.08
2025-01-30View details →
Mistral

Mistral: Pixtral Large 2411

Pixtral Large is a 124B parameter, open-weight, multimodal model built on top of [Mistral Large 2](/mistralai/mistral-large-2411). The model is able to understand documents, charts and natural images. The model is available under the Mistral Research License (MRL) for research and educational use, and the Mistral Commercial License for experimentation, testing, and production for commercial purposes.

Context
131K
Max Output
—
Input/1M
$2.00
👁 Vision🔧 Tools
Pricing (per 1M tokens)
OpenRouter$2.00 / $6.00
Martian$2.00 / $6.00
2024-11-19View details →
Mistral

Pixtral Large Latest

 

Context
131K
Max Output
—
Input/1M
$2.00
🔧 Tools
Pricing (per 1M tokens)
Requesty★$2.00 / $5.00
Vercel AI$2.00 / $6.00
2024-11-01View details →
Mistral

Ministral 3B

A compact, efficient model for on-device tasks like smart assistants and local analytics, offering low-latency performance.

Context
128K
Max Output
4K
Input/1M
$0.04
🔧 Tools
Pricing (per 1M tokens)
Vercel AI$0.04 / $0.04
2024-10-01View details →
MistralMistral Small

Mistral Small Latest

 

Context
33K
Max Output
—
Input/1M
$0.10
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.10 / $0.30
Vercel AI$0.10 / $0.30
2024-09-01View details →
Mistral

Pixtral 12B 2409

A 12B model with image understanding capabilities in addition to text.

Context
128K
Max Output
4K
Input/1M
$0.15
🔧 Tools
Pricing (per 1M tokens)
Vercel AI$0.15 / $0.15
2024-09-01View details →
Mistral

Mistral: Mistral Nemo

A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA. The model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi. It supports function calling and is released under the Apache 2.0 license.

Context
131K
Max Output
16K
Input/1M
$0.02
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.02 / $0.04
Vercel AI$0.02 / $0.04
Martian$0.02 / $0.04
2024-07-19View details →
Mistral

Codestral Latest

 

Context
131K
Max Output
—
Input/1M
$0.30
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.30 / $0.90
Vercel AI$0.30 / $0.90
2024-05-29View details →
MistralOSS

Mistral: Mistral 7B Instruct v0.3

A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length. An improved version of [Mistral 7B Instruct v0.2](/models/mistralai/mistral-7b-instruct-v0.2), with the following changes: - Extended vocabulary to 32768 - Supports v3 Tokenizer - Supports function calling NOTE: Support for function calling depends on the provider.

Context
33K
Max Output
4K
Input/1M
$0.20
Pricing (per 1M tokens)
OpenRouter$0.20 / $0.20
Martian$0.20 / $0.20
2024-05-27View details →
MistralMistral Large

Mistral Large Latest

This is Mistral AI's flagship model, Mistral Large 2 (version `mistral-large-2407`). It's a proprietary weights-available model and excels at reasoning, code, JSON, chat, and more. Read the launch announcement [here](https://mistral.ai/news/mistral-large-2407/). It supports dozens of languages including French, German, Spanish, Italian, Portuguese, Arabic, Hindi, Russian, Chinese, Japanese, and Korean, along with 80+ coding languages including Python, Java, C, C++, JavaScript, and Bash. Its long context window allows precise information recall from large documents.

Context
131K
Max Output
—
Input/1M
$2.00
🔧 Tools
Pricing (per 1M tokens)
Requesty★$2.00 / $6.00
OpenRouter$2.00 / $6.00
Martian$2.00 / $6.00
2024-02-26View details →
MistralOSS

Mistral: Mistral 7B Instruct v0.2

A high-performing, industry-standard 7.3B parameter model, with optimizations for speed and context length. An improved version of [Mistral 7B Instruct](/modelsmistralai/mistral-7b-instruct-v0.1), with the following changes: - 32k context window (vs 8k context in v0.1) - Rope-theta = 1e6 - No Sliding-Window Attention

Context
33K
Max Output
—
Input/1M
$0.20
Pricing (per 1M tokens)
OpenRouter$0.20 / $0.20
Martian$0.20 / $0.20
2023-12-28View details →
Mistral

Mistral Embed

General-purpose text embedding model for semantic search, similarity, clustering, and RAG workflows.

Context
0
Max Output
—
Input/1M
$0.10
Pricing (per 1M tokens)
Vercel AI$0.10 / Free
2023-12-11View details →
MistralOSS

Mistral: Mistral 7B Instruct v0.1

A 7.3B parameter model that outperforms Llama 2 13B on all benchmarks, with optimizations for speed and context length.

Context
3K
Max Output
—
Input/1M
$0.11
Pricing (per 1M tokens)
OpenRouter$0.11 / $0.19
Martian$0.11 / $0.19
2023-09-28View details →
Mistral

Devstral Latest

An enterprise grade text model, that excels at using tools to explore codebases, editing multiple files and power software engineering agents.

Context
256K
Max Output
—
Input/1M
Free
🔧 Tools
Pricing (per 1M tokens)
Requesty★Free / Free
View details →
Mistral

Devstral Medium 2507

 

Context
131K
Max Output
—
Input/1M
$0.40
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.40 / $2.00
View details →
MistralOSS

Open Mistral 7b

 

Context
33K
Max Output
—
Input/1M
$0.25
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.25 / $0.25
View details →
Mistral

Devstral Small 2507

 

Context
131K
Max Output
—
Input/1M
$0.10
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.10 / $0.30
View details →
Mistral

Mistral Nemo

A 12B parameter model with a 128k token context length built by Mistral in collaboration with NVIDIA. The model is multilingual, supporting English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Korean, Arabic, and Hindi. It supports function calling and is released under the Apache 2.0 license.

Context
131K
Max Output
—
Input/1M
$0.17
Pricing (per 1M tokens)
Requesty★$0.17 / $0.17
View details →
← Back to all providers