MiniMax: MiniMax M2

Text input Text output Free Option
Author's Description

MiniMax-M2 is a compact, high-efficiency large language model optimized for end-to-end coding and agentic workflows. With 10 billion activated parameters (230 billion total), it delivers near-frontier intelligence across general reasoning, tool use, and multi-step task execution while maintaining low latency and deployment efficiency. The model excels in code generation, multi-file editing, compile-run-fix loops, and test-validated repair, showing strong results on SWE-Bench Verified, Multi-SWE-Bench, and Terminal-Bench. It also performs competitively in agentic evaluations such as BrowseComp and GAIA, effectively handling long-horizon planning, retrieval, and recovery from execution errors. Benchmarked by [Artificial Analysis](https://artificialanalysis.ai/models/minimax-m2), MiniMax-M2 ranks among the top open-source models for composite intelligence, spanning mathematics, science, and instruction-following. Its small activation footprint enables fast inference, high concurrency, and improved unit economics, making it well-suited for large-scale agents, developer assistants, and reasoning-driven applications that require responsiveness and cost efficiency. To avoid degrading this model's performance, MiniMax highly recommends preserving reasoning between turns. Learn more about using reasoning_details to pass back reasoning in our [docs](https://openrouter.ai/docs/use-cases/reasoning-tokens#preserving-reasoning-blocks).

Key Specifications
Cost
$$$$$
Context
204K
Parameters
230B (Rumoured)
Released
Oct 23, 2025
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Reasoning Min P Stop Presence Penalty Logit Bias Seed Frequency Penalty Temperature Top P Max Tokens Include Reasoning
Features

This model supports the following features:

Reasoning
Performance Summary

MiniMax-M2, a compact 10-billion parameter model with 230 billion total, demonstrates near-frontier intelligence, particularly excelling in coding and agentic workflows. While its speed ranking places it among models with longer response times (10th percentile), its pricing is moderate (23rd percentile). A significant strength is its exceptional reliability, boasting a 98% success rate, indicating minimal technical failures. The model shows outstanding performance in coding benchmarks (93.9% accuracy, 90th percentile) and strong reasoning capabilities (96.0% accuracy, 87th percentile). It also performs well in General Knowledge (99.0% accuracy, 66th percentile). However, MiniMax-M2 exhibits notable weaknesses in Instruction Following (11.6% accuracy, 24th percentile) and Hallucinations (80.0% accuracy, 24th percentile), suggesting areas for improvement in adhering to complex directives and acknowledging uncertainty. Its performance in Mathematics (84.0% accuracy, 48th percentile) and Ethics (98.0% accuracy, 40th percentile) is competitive, while Email Classification is average (97.0% accuracy, 46th percentile). Its small activation footprint supports fast inference and high concurrency, making it suitable for responsive, cost-efficient applications despite its overall slower response times.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.3
Completion $1.2

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
Parasail
Parasail | minimax/minimax-m2 196K $0.25 / 1M tokens $1 / 1M tokens
Chutes
Chutes | minimax/minimax-m2 196K $0.15 / 1M tokens $0.45 / 1M tokens
AtlasCloud
AtlasCloud | minimax/minimax-m2 196K $0.28 / 1M tokens $1.15 / 1M tokens
Novita
Novita | minimax/minimax-m2 204K $0.3 / 1M tokens $1.2 / 1M tokens
Fireworks
Fireworks | minimax/minimax-m2 204K $0.3 / 1M tokens $1.2 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Strategy Free Executions Accuracy Cost Duration
Other Models by minimax