DeepSeek: R1 Distill Qwen 14B

Text input Text output Free Option
Author's Description

DeepSeek R1 Distill Qwen 14B is a distilled large language model based on [Qwen 2.5 14B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-14B), using outputs from [DeepSeek R1](/deepseek/deepseek-r1). It outperforms OpenAI's o1-mini across various benchmarks, achieving new state-of-the-art results for dense models. Other benchmark results include: - AIME 2024 pass@1: 69.7 - MATH-500 pass@1: 93.9 - CodeForces Rating: 1481 The model leverages fine-tuning from DeepSeek R1's outputs, enabling competitive performance comparable to larger frontier models.

Key Specifications
Cost
$$$
Context
64K
Parameters
14B
Released
Jan 29, 2025
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Stop Top P Seed Min P Frequency Penalty Max Tokens Reasoning Presence Penalty Include Reasoning Logit Bias Temperature
Features

This model supports the following features:

Reasoning
Performance Summary

DeepSeek R1 Distill Qwen 14B, created on January 29, 2025, demonstrates strong overall performance for a distilled model. It consistently ranks among the fastest models, achieving an Infinityth percentile in speed across 7 benchmarks, and offers competitive pricing, ranking in the 59th percentile across 6 benchmarks. The model exhibits high reliability with a 91% success rate, indicating consistent and usable responses. In terms of specific benchmarks, the model shows exceptional strength in Coding, achieving 93.0% accuracy (91st percentile), and strong Reasoning capabilities with 86.0% accuracy (85th percentile). Its Email Classification also performs well at 93.0% accuracy, though this places it in the 26th percentile. While its Ethics performance is respectable at 87.5% accuracy, it falls into the 23rd percentile. A notable weakness is observed in one instance of Instruction Following, showing 0.0% accuracy, though another run of the same benchmark yielded 44.0% accuracy. General Knowledge also presents a challenge, with 77.5% accuracy placing it in the 24th percentile. Despite some lower percentile rankings, the raw accuracy scores for Ethics, Email Classification, and General Knowledge are still commendable. The model's ability to leverage fine-tuning from DeepSeek R1's outputs allows it to achieve competitive performance comparable to larger frontier models, particularly in specialized areas like AIME and MATH, and competitive CodeForces ratings.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.15
Completion $0.15

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
Novita
Novita | deepseek/deepseek-r1-distill-qwen-14b 64K $0.15 / 1M tokens $0.15 / 1M tokens
GMICloud
GMICloud | deepseek/deepseek-r1-distill-qwen-14b 131K $0.15 / 1M tokens $0.15 / 1M tokens
Together
Together | deepseek/deepseek-r1-distill-qwen-14b 131K $1.6 / 1M tokens $1.6 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Free Executions Accuracy Cost Duration
Other Models by deepseek