OpenAI: GPT-4.1 Nano

Image input File input Text input Text output
Author's Description

For tasks that demand low latency, GPT‑4.1 nano is the fastest and cheapest model in the GPT-4.1 series. It delivers exceptional performance at a small size with its 1 million token context window, and scores 80.1% on MMLU, 50.3% on GPQA, and 9.8% on Aider polyglot coding – even higher than GPT‑4o mini. It’s ideal for tasks like classification or autocompletion.

Key Specifications
Cost
$$
Context
1M
Parameters
50B (Rumoured)
Released
Apr 14, 2025
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Logit Bias Tool Choice Seed Top P Top Logprobs Temperature Response Format Logprobs Max Tokens Presence Penalty Structured Outputs Tools Frequency Penalty Stop
Features

This model supports the following features:

Structured Outputs Response Format Tools
Performance Summary

OpenAI's GPT-4.1 Nano, created April 14, 2025, is positioned as a fast and cost-effective model within the GPT-4.1 series, ideal for low-latency tasks like classification and autocompletion. It performs among the fastest models, ranking in the 76th percentile for speed across six benchmarks, and offers competitive pricing, ranking in the 78th percentile for cost-effectiveness. Demonstrating exceptional reliability, it consistently provides usable responses with minimal technical failures, achieving a perfect 100th percentile. The model exhibits strong performance in specific areas. It achieved perfect accuracy in the Ethics (Baseline) benchmark, notably being the most accurate model at its price point and among models of similar speed. Its General Knowledge (Baseline) performance was also very strong at 96.5% accuracy. While its Coding (Baseline) accuracy was 84.0%, it was highlighted as the most accurate among models of comparable speed. However, its performance in Instruction Following (Baseline) and Reasoning (Baseline) was moderate at 56.0% accuracy for both, suggesting these areas could be potential weaknesses. The Email Classification (Baseline) accuracy of 92.0% was solid but ranked lower compared to other models. Its 1 million token context window and MMLU, GPQA, and Aider polyglot coding scores further underscore its capabilities for demanding tasks.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.1
Completion $0.4
Input Cache Read $0.025

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
OpenAI
OpenAI | openai/gpt-4.1-nano-2025-04-14 1M $0.1 / 1M tokens $0.4 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Free Executions Accuracy Cost Duration
Other Models by openai