Author's Description
DeepSeek R1 Distill Qwen 32B is a distilled large language model based on [Qwen 2.5 32B](https://huggingface.co/Qwen/Qwen2.5-32B), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models.\n\nOther benchmark results include:\n\n- AIME 2024 pass@1: 72.6\n- MATH-500 pass@1: 94.3\n- CodeForces Rating: 1691\n\nThe model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.
Key Specifications
Supported Parameters
This model supports the following parameters:
Features
This model supports the following features:
Performance Summary
DeepSeek: R1 Distill Qwen 32B demonstrates a compelling balance of performance and cost-effectiveness, positioning it as a strong contender in the distilled LLM space. While its speed ranking places it in the 13th percentile, indicating longer response times compared to many peers, it compensates significantly with its cost-effectiveness, ranking in the 66th percentile for price. This suggests it typically provides more economical solutions. The model exhibits exceptional capabilities in specific domains. It achieved a remarkable 93.0% accuracy in the Coding (Baseline) benchmark, placing it in the 94th percentile and notably being the most accurate model at its price point. Similarly, in Reasoning (Baseline), it scored 94.0% accuracy (89th percentile), again standing out as the most accurate model for its cost. General Knowledge also shows strong performance at 98.5% accuracy (69th percentile). However, its performance in Ethics (Baseline) and Email Classification (Baseline) is more moderate, with accuracies of 97.5% (40th percentile) and 96.0% (42nd percentile) respectively. Overall, DeepSeek: R1 Distill Qwen 32B's key strength lies in its high accuracy in complex problem-solving and code generation at a competitive price, making it a valuable option for applications where throughput is less critical than precision and cost efficiency. Its primary weakness is its slower processing speed.
Model Pricing
Current Pricing
Feature | Price (per 1M tokens) |
---|---|
Prompt | $0.075 |
Completion | $0.15 |
Price History
Available Endpoints
Provider | Endpoint Name | Context Length | Pricing (Input) | Pricing (Output) |
---|---|---|---|---|
DeepInfra
|
DeepInfra | deepseek/deepseek-r1-distill-qwen-32b | 131K | $0.075 / 1M tokens | $0.15 / 1M tokens |
Novita
|
Novita | deepseek/deepseek-r1-distill-qwen-32b | 64K | $0.075 / 1M tokens | $0.15 / 1M tokens |
GMICloud
|
GMICloud | deepseek/deepseek-r1-distill-qwen-32b | 131K | $0.075 / 1M tokens | $0.15 / 1M tokens |
Cloudflare
|
Cloudflare | deepseek/deepseek-r1-distill-qwen-32b | 80K | $0.075 / 1M tokens | $0.15 / 1M tokens |
Nineteen
|
Nineteen | deepseek/deepseek-r1-distill-qwen-32b | 16K | $0.075 / 1M tokens | $0.15 / 1M tokens |
Benchmark Results
Benchmark | Category | Reasoning | Free | Executions | Accuracy | Cost | Duration |
---|
Other Models by deepseek
|
Released | Params | Context |
|
Speed | Ability | Cost |
---|---|---|---|---|---|---|---|
DeepSeek: R1 Distill Qwen 7B | May 30, 2025 | 7B | 131K |
Text input
Text output
|
★ | ★ | $$$$ |
DeepSeek: Deepseek R1 0528 Qwen3 8B | May 29, 2025 | 8B | 131K |
Text input
Text output
|
★ | ★★★★★ | $$$ |
DeepSeek: R1 0528 | May 28, 2025 | ~671B | 128K |
Text input
Text output
|
★ | ★★★★★ | $$$$$ |
DeepSeek: DeepSeek Prover V2 | Apr 30, 2025 | ~671B | 131K |
Text input
Text output
|
★★★★ | ★★★★★ | $$$$ |
DeepSeek: DeepSeek V3 0324 | Mar 24, 2025 | ~685B | 163K |
Text input
Text output
|
★★★ | ★★★★★ | $$$ |
DeepSeek: R1 Distill Llama 8B | Feb 07, 2025 | 8B | 32K |
Text input
Text output
|
★ | ★★★ | $$ |
DeepSeek: R1 Distill Qwen 1.5B | Jan 31, 2025 | 5B | 131K |
Text input
Text output
|
★★★ | ★ | $$$ |
DeepSeek: R1 Distill Qwen 14B | Jan 29, 2025 | 14B | 64K |
Text input
Text output
|
★ | ★★★ | $$$ |
DeepSeek: R1 Distill Llama 70B | Jan 23, 2025 | 70B | 131K |
Text input
Text output
|
★ | ★★★★★ | $$$$ |
DeepSeek: R1 | Jan 20, 2025 | ~671B | 128K |
Text input
Text output
|
★★ | ★★★★ | $$$$ |
DeepSeek: DeepSeek V3 | Dec 26, 2024 | — | 163K |
Text input
Text output
|
★★★ | ★★★★ | $$$ |