OpenAI: GPT-4o-mini (2024-07-18)

Image input File input Text input Text output
Author's Description

GPT-4o mini is OpenAI's newest model after [GPT-4 Omni](/models/openai/gpt-4o), supporting both text and image inputs with text outputs. As their most advanced small model, it is many multiples more affordable than other recent frontier models, and more than 60% cheaper than [GPT-3.5 Turbo](/models/openai/gpt-3.5-turbo). It maintains SOTA intelligence, while being significantly more cost-effective. GPT-4o mini achieves an 82% score on MMLU and presently ranks higher than GPT-4 on chat preferences [common leaderboards](https://arena.lmsys.org/). Check out the [launch announcement](https://openai.com/index/gpt-4o-mini-advancing-cost-efficient-intelligence/) to learn more. #multimodal

Key Specifications
Cost
$$
Context
128K
Parameters
8B (Rumoured)
Released
Jul 17, 2024
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Top Logprobs Stop Structured Outputs Logprobs Presence Penalty Frequency Penalty Top P Max Tokens Tool Choice Response Format Logit Bias Seed Temperature Tools
Features

This model supports the following features:

Response Format Tools Structured Outputs
Performance Summary

OpenAI's GPT-4o-mini (2024-07-18) demonstrates a strong overall performance profile, particularly excelling in cost-efficiency and reliability. The model performs among the fastest models, ranking in the 80th percentile for speed across benchmarks. It offers highly competitive pricing, placing in the 74th percentile for cost-effectiveness. Notably, GPT-4o-mini exhibits exceptional reliability with a 100% success rate across all evaluated benchmarks, indicating consistent and usable responses. In terms of specific benchmark performance, GPT-4o-mini achieved perfect accuracy in Ethics (100%), making it the most accurate model at its price point and among models of similar speed. It also showed strong performance in General Knowledge (99%) and Email Classification (98%), indicating robust understanding and categorization capabilities. Its Coding accuracy (87%) is commendable. However, the model's performance in Hallucinations (84% accuracy) suggests room for improvement in acknowledging uncertainty, as it ranks in the lower 26th percentile for this specific metric. Mathematics (76%) and Reasoning (60%) also present areas where further development could enhance its capabilities. Despite these areas for improvement, GPT-4o-mini's combination of high reliability, speed, and cost-effectiveness, coupled with its strong performance in critical areas like ethics and general knowledge, positions it as a highly valuable and accessible AI model.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.15
Completion $0.6
Input Cache Read $0.075

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
OpenAI
OpenAI | openai/gpt-4o-mini-2024-07-18 128K $0.15 / 1M tokens $0.6 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Strategy Free Executions Accuracy Cost Duration
Other Models by openai