OpenAI: o3 Mini

Text input Text output
Author's Description

OpenAI o3-mini is a cost-efficient language model optimized for STEM reasoning tasks, particularly excelling in science, mathematics, and coding. This model supports the `reasoning_effort` parameter, which can be set to "high", "medium", or "low" to control the thinking time of the model. The default is "medium". OpenRouter also offers the model slug `openai/o3-mini-high` to default the parameter to "high". The model features three adjustable reasoning effort levels and supports key developer capabilities including function calling, structured outputs, and streaming, though it does not include vision processing capabilities. The model demonstrates significant improvements over its predecessor, with expert testers preferring its responses 56% of the time and noting a 39% reduction in major errors on complex questions. With medium reasoning effort settings, o3-mini matches the performance of the larger o1 model on challenging reasoning evaluations like AIME and GPQA, while maintaining lower latency and cost.

Key Specifications
Cost
$$$$$
Context
200K
Parameters
200B
Released
Jan 31, 2025
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Tool Choice Max Tokens Structured Outputs Tools Seed Response Format
Features

This model supports the following features:

Structured Outputs Response Format Tools
Performance Summary

OpenAI o3-mini, released on January 31, 2025, is a cost-efficient language model specifically optimized for STEM reasoning tasks. It demonstrates moderate speed performance, ranking in the 38th percentile across benchmarks, and offers moderate pricing, placing it in the 22nd percentile. A standout feature is its exceptional reliability, achieving the 100th percentile with minimal technical failures. The model excels in specialized areas, achieving 93.0% accuracy in Coding and 98.0% in Reasoning, placing it in the 91st and 95th percentiles respectively. This aligns with its design for STEM tasks, where it matches the performance of the larger o1 model on challenging evaluations like AIME and GPQA with medium reasoning effort. General Knowledge and Ethics also show strong accuracy at 99.5% and 99.0%. While its Instruction Following accuracy is moderate at 59.6%, its Email Classification is high at 98.0%. The model's ability to adjust reasoning effort ("high", "medium", "low") provides flexibility for balancing performance and cost. Key developer capabilities include function calling, structured outputs, and streaming, though it lacks vision processing. Expert testers preferred o3-mini's responses 56% of the time over its predecessor, noting a 39% reduction in major errors on complex questions.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $1.1
Completion $4.4
Input Cache Read $0.55

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
OpenAI
OpenAI | openai/o3-mini-2025-01-31 200K $1.1 / 1M tokens $4.4 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Free Executions Accuracy Cost Duration
Other Models by openai