LLM Router
HomeRoutersModelsProvidersBenchmarksPricingCompareBlogAbout
HomeModelsBenchmarksPricingCompareBlog
LLM Router

Independent comparison platform for LLM routing infrastructure.

Platform

  • Home
  • Routers
  • Models
  • Pricing
  • Blog
  • About

Routers

  • Requesty
  • OpenRouter
  • Martian
  • Unify
  • LiteLLM

© 2026 LLM Router

Data from public sources. May not reflect real-time pricing.

Providers›DeepSeek

DeepSeek

↗ Website

DeepSeek is a Chinese AI research lab known for highly capable open-source models. Their DeepSeek V3 and R1 reasoning models have achieved remarkable performance at significantly lower costs, challenging the dominance of proprietary models from Western labs.

Pricing available from Requesty, OpenRouter, Martian, Vercel AI, DeepInfra.

Total Models
29
Arena Ranked
8
of 29
Open Source
29
of 29
Cheapest Input
$0.15
per 1M tokens

$ Pricing Summary(per 1M tokens)

MetricInputOutput
Cheapest$0.15$0.15
Average$0.52$1.26
Most Expensive$4.00$7.00

⚙ Capabilities

👁
Vision
0
of 29 models
🧠
Reasoning
13
of 29 models
🔧
Tool Calling
22
of 29 models
⚡
Prompt Caching
8
of 29 models
🖥
Computer Use
1
of 29 models
🎨
Image Generation
0
of 29 models

🤖 All DeepSeek Models(29)

DeepSeekDeepSeek V3.2OSS
#37

DeepSeek: DeepSeek V3.2 Exp

DeepSeek-V3.2-Exp is an experimental large language model released by DeepSeek as an intermediate step between V3.1 and future architectures. It introduces DeepSeek Sparse Attention (DSA), a fine-grained sparse attention mechanism designed to improve training and inference efficiency in long-context scenarios while maintaining output quality. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config) The model was trained under conditions aligned with V3.1-Terminus to enable direct comparison. Benchmarking shows performance roughly on par with V3.1 across reasoning, coding, and agentic tool-use tasks, with minor tradeoffs and gains depending on the domain. This release focuses on validating architectural optimizations for extended context lengths rather than advancing raw task accuracy, making it primarily a research-oriented model for exploring efficient transformer designs.

Context
164K
Max Output
66K
Input/1M
$0.27
🧠 Reasoning🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.27 / $0.41
Martian$0.27 / $0.41
2025-09-29View details →
DeepSeekDeepSeek V3.2OSS
#40

DeepSeek: DeepSeek V3.2

DeepSeek-V3.2 is a large language model designed to harmonize high computational efficiency with strong reasoning and agentic tool-use performance. It introduces DeepSeek Sparse Attention (DSA), a fine-grained sparse attention mechanism that reduces training and inference cost while preserving quality in long-context scenarios. A scalable reinforcement learning post-training framework further improves reasoning, with reported performance in the GPT-5 class, and the model has demonstrated gold-medal results on the 2025 IMO and IOI. V3.2 also uses a large-scale agentic task synthesis pipeline to better integrate reasoning into tool-use settings, boosting compliance and generalization in interactive environments. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config)

Context
164K
Max Output
66K
Input/1M
$0.25
🧠 Reasoning🔧 Tools⚡ Cache
Pricing (per 1M tokens)
OpenRouter$0.25 / $0.38
Vercel AI$0.26 / $0.38
Martian$0.25 / $0.38
DeepInfra$0.26 / $0.38
2025-12-01View details →
DeepSeekDeepSeek V3.2OSS
#41

DeepSeek V3.2 Thinking

Thinking mode of DeepSeek V3.2

Context
128K
Max Output
64K
Input/1M
$0.28
🧠 Reasoning🔧 Tools
Pricing (per 1M tokens)
Vercel AI$0.28 / $0.42
2025-12-01View details →
DeepSeekDeepSeek R1OSS
#42

DeepSeek: R1 0528 (free)

May 28th update to the [original DeepSeek R1](/deepseek/deepseek-r1) Performance on par with [OpenAI o1](/openai/o1), but open-sourced and with fully open reasoning tokens. It's 671B parameters in size, with 37B active in an inference pass. Fully open-source model.

Context
164K
Max Output
164K
Input/1M
Free
🧠 Reasoning
Pricing (per 1M tokens)
OpenRouterFree / Free
Martian$0.40 / $1.75
DeepInfra$0.50 / $2.15
2025-05-28View details →
DeepSeekDeepSeek V3.1OSS
#44

DeepSeek-V3.1

DeepSeek-V3.1 is post-trained on the top of DeepSeek-V3.1-Base, which is built upon the original V3 base checkpoint through a two-phase long context extension approach, following the methodology outlined in the original DeepSeek-V3 report. We have expanded our dataset by collecting additional long documents and substantially extending both training phases. The 32K extension phase has been increased 10-fold to 630B tokens, while the 128K extension phase has been extended by 3.3x to 209B tokens. Additionally, DeepSeek-V3.1 is trained using the UE8M0 FP8 scale data format to ensure compatibility with microscaling data formats.

Context
164K
Max Output
128K
Input/1M
$0.21
🧠 Reasoning🔧 Tools
Pricing (per 1M tokens)
Vercel AI$0.21 / $0.79
DeepInfra$0.21 / $0.79
2025-08-21View details →
DeepSeekDeepSeek V3.1OSS
#49

DeepSeek: DeepSeek V3.1 Terminus (exacto)

DeepSeek-V3.1 Terminus is an update to [DeepSeek V3.1](/deepseek/deepseek-chat-v3.1) that maintains the model's original capabilities while addressing issues reported by users, including language consistency and agent capabilities, further optimizing the model's performance in coding and search agents. It is a large hybrid reasoning model (671B parameters, 37B active) that supports both thinking and non-thinking modes. It extends the DeepSeek-V3 base with a two-phase long-context training process, reaching up to 128K tokens, and uses FP8 microscaling for efficient inference. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config) The model improves tool use, code generation, and reasoning efficiency, achieving performance comparable to DeepSeek-R1 on difficult benchmarks while responding more quickly. It supports structured tool calling, code agents, and search agents, making it suitable for research, coding, and agentic workflows.

Context
164K
Max Output
—
Input/1M
$0.21
🧠 Reasoning🔧 Tools⚡ Cache
Pricing (per 1M tokens)
OpenRouter$0.21 / $0.79
Vercel AI$0.27 / $1.00
Martian$0.21 / $0.79
DeepInfra$0.21 / $0.79
2025-09-22View details →
DeepSeekDeepSeek R1OSS
#68

DeepSeek: R1

DeepSeek R1 is here: Performance on par with [OpenAI o1](/openai/o1), but open-sourced and with fully open reasoning tokens. It's 671B parameters in size, with 37B active in an inference pass. Fully open-source model & [technical report](https://api-docs.deepseek.com/news/news250120). MIT licensed: Distill & commercialize freely!

Context
64K
Max Output
16K
Input/1M
$0.50
🧠 Reasoning🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.70 / $2.50
Vercel AI$0.50 / $2.15
Martian$0.70 / $2.50
2025-01-20View details →
DeepSeekDeepSeek V3OSS
#104

DeepSeek V3 0324

Fast general-purpose LLM with enhanced reasoning capabilities

Context
164K
Max Output
16K
Input/1M
$0.32
🔧 Tools
Pricing (per 1M tokens)
Vercel AI$0.77 / $0.77
DeepInfra$0.32 / $0.89
2024-12-26View details →
DeepSeekDeepSeek V3.1OSS

Nex AGI: DeepSeek V3.1 Nex N1

DeepSeek V3.1 Nex-N1 is the flagship release of the Nex-N1 series — a post-trained model designed to highlight agent autonomy, tool use, and real-world productivity. Nex-N1 demonstrates competitive performance across all evaluation scenarios, showing particularly strong results in practical coding and HTML generation tasks.

Context
131K
Max Output
164K
Input/1M
$0.27
🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.27 / $1.00
Martian$0.27 / $1.00
2025-12-08View details →
DeepSeekDeepSeek V3.2OSS

DeepSeek: DeepSeek V3.2 Speciale

DeepSeek-V3.2-Speciale is a high-compute variant of DeepSeek-V3.2 optimized for maximum reasoning and agentic performance. It builds on DeepSeek Sparse Attention (DSA) for efficient long-context processing, then scales post-training reinforcement learning to push capability beyond the base model. Reported evaluations place Speciale ahead of GPT-5 on difficult reasoning workloads, with proficiency comparable to Gemini-3.0-Pro, while retaining strong coding and tool-use reliability. Like V3.2, it benefits from a large-scale agentic task synthesis pipeline that improves compliance and generalization in interactive environments.

Context
164K
Max Output
66K
Input/1M
$0.27
🧠 Reasoning⚡ Cache
Pricing (per 1M tokens)
OpenRouter$0.27 / $0.41
Martian$0.27 / $0.41
2025-12-01View details →
DeepSeekOSS

DeepSeek: DeepSeek V3.1

DeepSeek-V3.1 is a large hybrid reasoning model (671B parameters, 37B active) that supports both thinking and non-thinking modes via prompt templates. It extends the DeepSeek-V3 base with a two-phase long-context training process, reaching up to 128K tokens, and uses FP8 microscaling for efficient inference. Users can control the reasoning behaviour with the `reasoning` `enabled` boolean. [Learn more in our docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#enable-reasoning-with-default-config) The model improves tool use, code generation, and reasoning efficiency, achieving performance comparable to DeepSeek-R1 on difficult benchmarks while responding more quickly. It supports structured tool calling, code agents, and search agents, making it suitable for research, coding, and agentic workflows. It succeeds the [DeepSeek V3-0324](/deepseek/deepseek-chat-v3-0324) model and performs well on a variety of tasks.

Context
33K
Max Output
7K
Input/1M
$0.15
🧠 Reasoning🔧 Tools
Pricing (per 1M tokens)
OpenRouter$0.15 / $0.75
Martian$0.15 / $0.75
2025-08-21View details →
DeepSeekDeepSeek R1OSS

TNG: DeepSeek R1T2 Chimera

DeepSeek-TNG-R1T2-Chimera is the second-generation Chimera model from TNG Tech. It is a 671 B-parameter mixture-of-experts text-generation model assembled from DeepSeek-AI’s R1-0528, R1, and V3-0324 checkpoints with an Assembly-of-Experts merge. The tri-parent design yields strong reasoning performance while running roughly 20 % faster than the original R1 and more than 2× faster than R1-0528 under vLLM, giving a favorable cost-to-intelligence trade-off. The checkpoint supports contexts up to 60 k tokens in standard use (tested to ~130 k) and maintains consistent <think> token behaviour, making it suitable for long-context analysis, dialogue and other open-ended generation tasks.

Context
164K
Max Output
164K
Input/1M
$0.25
🧠 Reasoning🔧 Tools⚡ Cache
Pricing (per 1M tokens)
OpenRouter$0.25 / $0.85
2025-07-08View details →
DeepSeekDeepSeek R1OSS

TNG: DeepSeek R1T Chimera

DeepSeek-R1T-Chimera is created by merging DeepSeek-R1 and DeepSeek-V3 (0324), combining the reasoning capabilities of R1 with the token efficiency improvements of V3. It is based on a DeepSeek-MoE Transformer architecture and is optimized for general text generation tasks. The model merges pretrained weights from both source models to balance performance across reasoning, efficiency, and instruction-following tasks. It is released under the MIT license and intended for research and commercial use.

Context
164K
Max Output
164K
Input/1M
$0.30
🧠 Reasoning⚡ Cache
Pricing (per 1M tokens)
OpenRouter$0.30 / $1.20
2025-04-27View details →
DeepSeekOSS

DeepSeek: DeepSeek V3 0324

DeepSeek V3, a 685B-parameter, mixture-of-experts model, is the latest iteration of the flagship chat model family from the DeepSeek team. It succeeds the [DeepSeek V3](/deepseek/deepseek-chat-v3) model and performs really well on a variety of tasks.

Context
164K
Max Output
66K
Input/1M
$0.19
🧠 Reasoning🔧 Tools⚡ Cache
Pricing (per 1M tokens)
OpenRouter$0.19 / $0.87
Martian$0.19 / $0.87
2025-03-24View details →
DeepSeekDeepSeek R1OSS

DeepSeek: R1 Distill Qwen 32B

DeepSeek R1 Distill Qwen 32B is a distilled large language model based on [Qwen 2.5 32B](https://huggingface.co/Qwen/Qwen2.5-32B), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.\n\nOther benchmark results include:\n\n- AIME 2024 pass@1: 72.6\n- MATH-500 pass@1: 94.3\n- CodeForces Rating: 1691\n\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.

Context
33K
Max Output
33K
Input/1M
$0.29
🧠 Reasoning
Pricing (per 1M tokens)
OpenRouter$0.29 / $0.29
Martian$0.29 / $0.29
2025-01-29View details →
DeepSeekOSS

Deepseek Chat

DeepSeek-V3 is the latest model from the DeepSeek team, building upon the instruction following and coding abilities of the previous versions. Pre-trained on nearly 15 trillion tokens, the reported evaluations reveal that the model outperforms other open-source models and rivals leading closed-source models. For model details, please visit [the DeepSeek-V3 repo](https://github.com/deepseek-ai/DeepSeek-V3) for more information, or see the [launch announcement](https://api-docs.deepseek.com/news/news1226).

Context
164K
Max Output
164K
Input/1M
$0.28
🔧 Tools⚡ Cache
Pricing (per 1M tokens)
Requesty cheapest
Requesty★$0.28 / $0.42
OpenRouter$0.30 / $1.20
Martian$0.30 / $1.20
2024-12-26View details →
DeepSeekDeepSeek V3OSS

DeepSeek V3 0324

DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.

Context
128K
Max Output
—
Input/1M
$0.50
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.50 / $1.50
View details →
DeepSeekDeepSeek R1OSS

DeepSeek R1 0528

DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.

Context
164K
Max Output
—
Input/1M
$0.80
🔧 Tools🖥 Computer
Pricing (per 1M tokens)
Requesty★$0.80 / $2.40
View details →
DeepSeekDeepSeek V3OSS

DeepSeek V3 0324 Fast

DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.

Context
128K
Max Output
—
Input/1M
$2.00
Pricing (per 1M tokens)
Requesty★$2.00 / $6.00
View details →
DeepSeekDeepSeek R1OSS

Deepseek R1 Distill Qwen 14b

DeepSeek R1 Distill Qwen 14B is a distilled large language model based on Qwen 2.5 14B, using outputs from DeepSeek R1. It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models. Other benchmark results include: AIME 2024 pass@1: 69.7 MATH-500 pass@1: 93.9 CodeForces Rating: 1481 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.

Context
128K
Max Output
—
Input/1M
$0.15
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.15 / $0.15
View details →
DeepSeekOSS

Deepseek V3

DeepSeek-V3 is the latest model from the DeepSeek team, building upon the instruction following and coding abilities of the previous versions. Pre-trained on nearly 15 trillion tokens, the reported evaluations reveal that the model outperforms other open-source models and rivals leading closed-source models.

Context
64K
Max Output
—
Input/1M
$0.89
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.89 / $0.89
View details →
DeepSeekDeepSeek V3OSS

Deepseek V3 0324

DeepSeek R1 is the latest open-source model released by the DeepSeek team, featuring impressive reasoning capabilities, particularly achieving performance comparable to OpenAI's o1 model in mathematics, coding, and reasoning tasks.

Context
128K
Max Output
—
Input/1M
$0.40
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.40 / $1.30
View details →
DeepSeekDeepSeek R1OSS

DeepSeek R1

DeepSeek R1 is the latest open-source model released by the DeepSeek team, featuring impressive reasoning capabilities, particularly achieving performance comparable to OpenAI's o1 model in mathematics, coding, and reasoning tasks.

Context
64K
Max Output
—
Input/1M
$4.00
🔧 Tools
Pricing (per 1M tokens)
Requesty★$4.00 / $4.00
View details →
DeepSeekDeepSeek R1OSS

Deepseek R1 Distill Qwen 32b

DeepSeek R1 Distill Qwen 32B is a distilled large language model based on Qwen 2.5 32B, using outputs from DeepSeek R1. It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models. Other benchmark results include: AIME 2024 pass@1: 72.6 MATH-500 pass@1: 94.3 CodeForces Rating: 1691 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.

Context
13K
Max Output
—
Input/1M
$0.30
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.30 / $0.30
View details →
DeepSeekOSS

Deepseek Prover V2 671b

DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.

Context
160K
Max Output
—
Input/1M
$0.70
Pricing (per 1M tokens)
Requesty★$0.70 / $2.50
View details →
DeepSeekDeepSeek V3OSS

DeepSeek V3

DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.

Context
128K
Max Output
8K
Input/1M
$0.85
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.85 / $0.90
View details →
DeepSeekDeepSeek R1OSS

DeepSeek R1

DeepSeek-R1-Distill-Qwen-7B is a 7 billion parameter dense language model distilled from DeepSeek-R1, leveraging reinforcement learning-enhanced reasoning data generated by DeepSeek's larger models. The distillation process transfers advanced reasoning, math, and code capabilities into a smaller, more efficient model architecture based on Qwen2.5-Math-7B. This model demonstrates strong performance across mathematical benchmarks (92.8% pass@1 on MATH-500), coding tasks (Codeforces rating 1189), and general reasoning (49.1% pass@1 on GPQA Diamond), achieving competitive accuracy relative to larger models while maintaining smaller inference costs.

Context
64K
Max Output
8K
Input/1M
$3.00
Pricing (per 1M tokens)
Requesty★$3.00 / $7.00
View details →
DeepSeekOSS

Deepseek Reasoner

Fully open-source model & technical report. Performance on par with OpenAI-o1.

Context
128K
Max Output
64K
Input/1M
$0.28
🔧 Tools⚡ Cache
Pricing (per 1M tokens)
Requesty★$0.28 / $0.42
View details →
DeepSeekDeepSeek V3.1OSS

DeepSeek V3.1

 

Context
164K
Max Output
—
Input/1M
$0.30
🔧 Tools
Pricing (per 1M tokens)
Requesty★$0.30 / $1.00
View details →
← Back to all providers