Mistral: Mistral Small 3.2 24B

Text input Image input Text output
Author's Description

Mistral-Small-3.2-24B-Instruct-2506 is an updated 24B parameter model from Mistral optimized for instruction following, repetition reduction, and improved function calling. Compared to the 3.1 release, version 3.2 significantly improves accuracy on WildBench and Arena Hard, reduces infinite generations, and delivers gains in tool use and structured output tasks. It supports image and text inputs with structured outputs, function/tool calling, and strong performance across coding (HumanEval+, MBPP), STEM (MMLU, MATH, GPQA), and vision benchmarks (ChartQA, DocVQA).

Key Specifications
Cost
$$
Context
131K
Parameters
24B
Released
Jun 20, 2025
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Stop Presence Penalty Tool Choice Top P Temperature Seed Tools Structured Outputs Response Format Frequency Penalty Max Tokens
Features

This model supports the following features:

Tools Structured Outputs Response Format
Performance Summary

Mistral-Small-3.2-24B-Instruct-2506, created on June 20, 2025, demonstrates a strong overall performance profile, particularly excelling in reliability and cost-efficiency. It consistently offers among the most competitive pricing, ranking in the 82nd percentile, and exhibits exceptional reliability with a perfect 100th percentile ranking, indicating minimal technical failures. While its speed is competitive, ranking in the 59th percentile, it does not consistently perform among the fastest models. Across benchmark categories, the model shows notable strengths in classification and ethical reasoning. It achieved perfect accuracy in the Ethics (Baseline) benchmark, making it the most accurate model at its price point and among models of similar speed. Email Classification also saw high accuracy at 99.0%. Coding performance is solid at 86.0% accuracy. However, its performance in Instruction Following (51.0% accuracy) and Reasoning (56.0% accuracy) suggests areas for potential improvement, as these scores are closer to the median. General Knowledge is strong at 98.5%. The model's ability to handle image and text inputs with structured outputs and function/tool calling further enhances its versatility.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.1
Completion $0.3

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
Mistral
Mistral | mistralai/mistral-small-3.2-24b-instruct-2506 131K $0.1 / 1M tokens $0.3 / 1M tokens
DeepInfra
DeepInfra | mistralai/mistral-small-3.2-24b-instruct-2506 128K $0.05 / 1M tokens $0.1 / 1M tokens
Enfer
Enfer | mistralai/mistral-small-3.2-24b-instruct-2506 32K $0.075 / 1M tokens $0.2 / 1M tokens
Parasail
Parasail | mistralai/mistral-small-3.2-24b-instruct-2506 131K $0.09 / 1M tokens $0.3 / 1M tokens
Chutes
Chutes | mistralai/mistral-small-3.2-24b-instruct-2506 131K $0.104 / 1M tokens $0.415 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Free Executions Accuracy Cost Duration
Other Models by mistralai