Mistral: Mistral Medium 3.1

Image input Text input Text output
Author's Description

Mistral Medium 3.1 is an updated version of Mistral Medium 3, which is a high-performance enterprise-grade language model designed to deliver frontier-level capabilities at significantly reduced operational cost. It balances state-of-the-art reasoning and multimodal performance with 8× lower cost compared to traditional large models, making it suitable for scalable deployments across professional and industrial use cases. The model excels in domains such as coding, STEM reasoning, and enterprise adaptation. It supports hybrid, on-prem, and in-VPC deployments and is optimized for integration into custom workflows. Mistral Medium 3.1 offers competitive accuracy relative to larger models like Claude Sonnet 3.5/3.7, Llama 4 Maverick, and Command R+, while maintaining broad compatibility across cloud environments.

Key Specifications
Cost
$$$$
Context
131K
Released
Aug 13, 2025
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Response Format Presence Penalty Tools Top P Frequency Penalty Max Tokens Structured Outputs Seed Tool Choice Stop Temperature
Features

This model supports the following features:

Tools Structured Outputs Response Format
Performance Summary

Mistral Medium 3.1, created on August 13, 2025, is an enterprise-grade language model designed for high performance at reduced operational costs. It demonstrates competitive response times, ranking in the 59th percentile for speed, and offers competitive pricing, placing in the 43rd percentile. Notably, the model exhibits exceptional reliability with a 100% success rate across all benchmarks, indicating consistent and usable responses. The model excels in several key areas. It achieved perfect accuracy in both Hallucinations (Baseline) and General Knowledge (Baseline) tests, demonstrating robust factual recall and appropriate uncertainty acknowledgment. Its performance in Mathematics (93.0% accuracy) and Coding (92.0% accuracy) benchmarks is particularly strong, placing it in the 81st and 80th percentiles respectively, highlighting its capabilities in STEM and development tasks. While its Ethics (99.0%) and Email Classification (98.0%) scores are also high, its Instruction Following (66.0%) and Reasoning (72.0%) benchmarks, though respectable, indicate areas where further refinement could enhance its overall utility for complex, multi-step tasks. Mistral Medium 3.1's balance of state-of-the-art reasoning and multimodal performance, coupled with its cost-effectiveness and broad compatibility, positions it as a strong contender for scalable professional and industrial deployments.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.4
Completion $2

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
Mistral
Mistral | mistralai/mistral-medium-3.1 131K $0.4 / 1M tokens $2 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Strategy Free Executions Accuracy Cost Duration
Other Models by mistralai