Author's Description
DeepSeek V3, a 685B-parameter, mixture-of-experts model, is the latest iteration of the flagship chat model family from the DeepSeek team. It succeeds the [DeepSeek V3](/deepseek/deepseek-chat-v3) model and performs really well on a variety of tasks.
Key Specifications
Supported Parameters
This model supports the following parameters:
Features
This model supports the following features:
Performance Summary
DeepSeek V3 (0324), a 685B-parameter mixture-of-experts model, demonstrates a strong overall performance profile. It exhibits competitive response times, ranking in the 48th percentile across seven benchmarks, indicating it performs among moderately fast models. Its pricing is also competitive, placing it in the 59th percentile across six benchmarks. A standout feature is its exceptional reliability, achieving a perfect 100% success rate across all seven benchmarks, signifying consistent and dependable operation with minimal technical failures. Across benchmark categories, DeepSeek V3 shows particular strength in Instruction Following and Ethics, achieving perfect 100% accuracy in both. For Instruction Following, it is noted as the most accurate among models of comparable speed. In Ethics, it is highlighted as both the most accurate model at its price point and among models of similar speed. It also performs very well in General Knowledge (99.5% accuracy, 83rd percentile) and Coding (92.0% accuracy, 84th percentile). While its Reasoning accuracy is solid at 80.0% (77th percentile), and Email Classification at 97.0% (51st percentile), these are not as exceptionally high as its top categories. The model's key strengths lie in its high accuracy for critical tasks like instruction following and ethical reasoning, coupled with its robust reliability. No significant weaknesses are apparent from the provided data.
Model Pricing
Current Pricing
Feature | Price (per 1M tokens) |
---|---|
Prompt | $0.28 |
Completion | $0.88 |
Price History
Available Endpoints
Provider | Endpoint Name | Context Length | Pricing (Input) | Pricing (Output) |
---|---|---|---|---|
DeepInfra
|
DeepInfra | deepseek/deepseek-chat-v3-0324 | 163K | $0.28 / 1M tokens | $0.88 / 1M tokens |
Novita
|
Novita | deepseek/deepseek-chat-v3-0324 | 163K | $0.28 / 1M tokens | $1.14 / 1M tokens |
Kluster
|
Kluster | deepseek/deepseek-chat-v3-0324 | 163K | $0.2 / 1M tokens | $0.8 / 1M tokens |
Lambda
|
Lambda | deepseek/deepseek-chat-v3-0324 | 163K | $0.34 / 1M tokens | $0.88 / 1M tokens |
Atoma
|
Atoma | deepseek/deepseek-chat-v3-0324 | 100K | $0.2 / 1M tokens | $0.8 / 1M tokens |
Cent-ML
|
Cent-ML | deepseek/deepseek-chat-v3-0324 | 32K | $0.2 / 1M tokens | $0.8 / 1M tokens |
InferenceNet
|
InferenceNet | deepseek/deepseek-chat-v3-0324 | 128K | $0.2 / 1M tokens | $0.8 / 1M tokens |
Nebius
|
Nebius | deepseek/deepseek-chat-v3-0324 | 163K | $0.5 / 1M tokens | $1.5 / 1M tokens |
GMICloud
|
GMICloud | deepseek/deepseek-chat-v3-0324 | 163K | $0.49 / 1M tokens | $0.9 / 1M tokens |
BaseTen
|
BaseTen | deepseek/deepseek-chat-v3-0324 | 163K | $0.77 / 1M tokens | $0.77 / 1M tokens |
Parasail
|
Parasail | deepseek/deepseek-chat-v3-0324 | 163K | $0.79 / 1M tokens | $1.15 / 1M tokens |
Fireworks
|
Fireworks | deepseek/deepseek-chat-v3-0324 | 163K | $0.9 / 1M tokens | $0.9 / 1M tokens |
Hyperbolic
|
Hyperbolic | deepseek/deepseek-chat-v3-0324 | 163K | $1.25 / 1M tokens | $1.25 / 1M tokens |
Together
|
Together | deepseek/deepseek-chat-v3-0324 | 131K | $1.25 / 1M tokens | $1.25 / 1M tokens |
SambaNova
|
SambaNova | deepseek/deepseek-chat-v3-0324 | 32K | $3 / 1M tokens | $4.5 / 1M tokens |
DeepSeek
|
DeepSeek | deepseek/deepseek-chat-v3-0324 | 64K | $0.2 / 1M tokens | $0.8 / 1M tokens |
Crusoe
|
Crusoe | deepseek/deepseek-chat-v3-0324 | 131K | $0.2 / 1M tokens | $0.8 / 1M tokens |
Phala
|
Phala | deepseek/deepseek-chat-v3-0324 | 163K | $0.49 / 1M tokens | $1.14 / 1M tokens |
AtlasCloud
|
AtlasCloud | deepseek/deepseek-chat-v3-0324 | 131K | $0.32 / 1M tokens | $1.15 / 1M tokens |
Crusoe
|
Crusoe | deepseek/deepseek-chat-v3-0324 | 131K | $0.28 / 1M tokens | $1.15 / 1M tokens |
Targon
|
Targon | deepseek/deepseek-chat-v3-0324 | 163K | $0.25 / 1M tokens | $0.85 / 1M tokens |
Chutes
|
Chutes | deepseek/deepseek-chat-v3-0324 | 163K | $0.2 / 1M tokens | $0.8 / 1M tokens |
SiliconFlow
|
SiliconFlow | deepseek/deepseek-chat-v3-0324 | 163K | $0.27 / 1M tokens | $1.13 / 1M tokens |
Benchmark Results
Benchmark | Category | Reasoning | Free | Executions | Accuracy | Cost | Duration |
---|
Other Models by deepseek
|
Released | Params | Context |
|
Speed | Ability | Cost |
---|---|---|---|---|---|---|---|
DeepSeek: DeepSeek V3.1 | Aug 21, 2025 | ~671B | 131K |
Text input
Text output
|
★★ | ★★★★★ | $$$ |
DeepSeek: DeepSeek V3.1 Base | Aug 20, 2025 | ~671B | 163K |
Text input
Text output
|
★★ | ★ | $$ |
DeepSeek: R1 Distill Qwen 7B Unavailable | May 30, 2025 | 7B | 131K |
Text input
Text output
|
★ | ★ | $$$$ |
DeepSeek: Deepseek R1 0528 Qwen3 8B | May 29, 2025 | 8B | 131K |
Text input
Text output
|
★★★ | ★★★ | $$ |
DeepSeek: R1 0528 | May 28, 2025 | ~671B | 128K |
Text input
Text output
|
★★★ | ★★★ | $$$ |
DeepSeek: DeepSeek Prover V2 | Apr 30, 2025 | ~671B | 131K |
Text input
Text output
|
★★ | ★★★★★ | $$$$ |
DeepSeek: DeepSeek V3 Base Unavailable | Mar 29, 2025 | ~671B | 163K |
Text input
Text output
|
★ | ★ | $$$ |
DeepSeek: R1 Distill Llama 8B | Feb 07, 2025 | 8B | 32K |
Text input
Text output
|
★ | ★★ | $$ |
DeepSeek: R1 Distill Qwen 1.5B Unavailable | Jan 31, 2025 | 5B | 131K |
Text input
Text output
|
★★★ | ★ | $$$ |
DeepSeek: R1 Distill Qwen 32B | Jan 29, 2025 | 32B | 131K |
Text input
Text output
|
★ | ★★★★★ | $$$ |
DeepSeek: R1 Distill Qwen 14B | Jan 29, 2025 | 14B | 64K |
Text input
Text output
|
★ | ★★ | $$$ |
DeepSeek: R1 Distill Llama 70B | Jan 23, 2025 | 70B | 131K |
Text input
Text output
|
★★★ | ★★★★★ | $$ |
DeepSeek: R1 | Jan 20, 2025 | ~671B | 128K |
Text input
Text output
|
★★★ | ★★★★★ | $$$ |
DeepSeek: DeepSeek V3 | Dec 26, 2024 | — | 163K |
Text input
Text output
|
★★★ | ★★★★★ | $$$ |