Author's Description
Llama 3.2 3B is a 3-billion-parameter multilingual large language model, optimized for advanced natural language processing tasks like dialogue generation, reasoning, and summarization. Designed with the latest transformer architecture, it supports eight languages, including English, Spanish, and Hindi, and is adaptable for additional languages. Trained on 9 trillion tokens, the Llama 3.2 3B model excels in instruction-following, complex reasoning, and tool use. Its balanced performance makes it ideal for applications needing accuracy and efficiency in text generation across multilingual settings. Click here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md). Usage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).
Key Specifications
Supported Parameters
This model supports the following parameters:
Features
This model supports the following features:
Performance Summary
Meta's Llama 3.2 3B Instruct model demonstrates exceptional speed, consistently ranking among the fastest models available, and offers highly competitive pricing, placing it in the 98th percentile across benchmarks. This 3-billion-parameter multilingual model, optimized for advanced NLP tasks, was created on September 24, 2024, and supports eight languages. In terms of performance across benchmarks, Llama 3.2 3B shows a mixed profile. It exhibits a notable strength in handling uncertainty, achieving 78.0% accuracy in Hallucinations (Baseline) by appropriately acknowledging fictional concepts. It also performs well in Email Classification, with 88.0% accuracy, making it the most accurate model at its price point for this task. General Knowledge is fair at 76.0% accuracy, boasting the best accuracy-to-cost ratio. However, the model struggles significantly in more complex cognitive areas. Its accuracy is notably low in Mathematics (4.0%), Ethics (44.0%), Reasoning (20.4%), and Coding (11.0%). Instruction Following also presents a weakness, with one benchmark showing 0.0% accuracy and another at 26.3%. These results suggest that while the model excels in specific classification and uncertainty handling tasks, its capabilities in complex problem-solving, ethical reasoning, and precise instruction execution require further development.
Model Pricing
Current Pricing
Feature | Price (per 1M tokens) |
---|---|
Prompt | $0.02 |
Completion | $0.02 |
Price History
Available Endpoints
Provider | Endpoint Name | Context Length | Pricing (Input) | Pricing (Output) |
---|---|---|---|---|
DeepInfra
|
DeepInfra | meta-llama/llama-3.2-3b-instruct | 131K | $0.02 / 1M tokens | $0.02 / 1M tokens |
Lambda
|
Lambda | meta-llama/llama-3.2-3b-instruct | 131K | $0.02 / 1M tokens | $0.02 / 1M tokens |
InferenceNet
|
InferenceNet | meta-llama/llama-3.2-3b-instruct | 16K | $0.02 / 1M tokens | $0.02 / 1M tokens |
Novita
|
Novita | meta-llama/llama-3.2-3b-instruct | 32K | $0.03 / 1M tokens | $0.05 / 1M tokens |
Cloudflare
|
Cloudflare | meta-llama/llama-3.2-3b-instruct | 128K | $0.051 / 1M tokens | $0.34 / 1M tokens |
Together
|
Together | meta-llama/llama-3.2-3b-instruct | 131K | $0.06 / 1M tokens | $0.06 / 1M tokens |
SambaNova
|
SambaNova | meta-llama/llama-3.2-3b-instruct | 4K | $0.02 / 1M tokens | $0.02 / 1M tokens |
Hyperbolic
|
Hyperbolic | meta-llama/llama-3.2-3b-instruct | 131K | $0.1 / 1M tokens | $0.1 / 1M tokens |
Nineteen
|
Nineteen | meta-llama/llama-3.2-3b-instruct | 20K | $0.02 / 1M tokens | $0.02 / 1M tokens |
Benchmark Results
Benchmark | Category | Reasoning | Strategy | Free | Executions | Accuracy | Cost | Duration |
---|
Other Models by meta-llama
|
Released | Params | Context |
|
Speed | Ability | Cost |
---|---|---|---|---|---|---|---|
Meta: Llama Guard 4 12B | Apr 29, 2025 | 12B | 163K |
Text input
Image input
Text output
|
— | ★ | $$ |
Meta: Llama 4 Maverick | Apr 05, 2025 | 17B | 1M |
Text input
Image input
Text output
|
★★★★★ | ★★★ | $$$ |
Meta: Llama 4 Scout | Apr 05, 2025 | 17B | 327K |
Text input
Image input
Text output
|
★★★★ | ★★ | $$ |
Llama Guard 3 8B | Feb 12, 2025 | 8B | 131K |
Text input
Text output
|
★★ | ★ | $$ |
Meta: Llama 3.3 70B Instruct | Dec 06, 2024 | 70B | 131K |
Text input
Text output
|
★★★★ | ★★★★ | $ |
Meta: Llama 3.2 1B Instruct | Sep 24, 2024 | 1B | 131K |
Text input
Text output
|
★★ | ★ | $ |
Meta: Llama 3.2 11B Vision Instruct | Sep 24, 2024 | 11B | 128K |
Text input
Image input
Text output
|
★★ | ★★ | $$ |
Meta: Llama 3.2 90B Vision Instruct | Sep 24, 2024 | 90B | 131K |
Text input
Image input
Text output
|
★★★ | ★★ | $$$$ |
Meta: Llama 3.1 405B (base) | Aug 01, 2024 | 405B | 32K |
Text input
Text output
|
★ | ★ | $$$ |
Meta: Llama 3.1 70B Instruct | Jul 22, 2024 | 70B | 131K |
Text input
Text output
|
★★★★ | ★★ | $$ |
Meta: Llama 3.1 405B Instruct | Jul 22, 2024 | 405B | 32K |
Text input
Text output
|
★★★★ | ★★ | $$$ |
Meta: Llama 3.1 8B Instruct | Jul 22, 2024 | 8B | 131K |
Text input
Text output
|
★★★ | ★★ | $ |
Meta: LlamaGuard 2 8B | May 12, 2024 | 8B | 8K |
Text input
Text output
|
★★★★ | ★ | $$ |
Meta: Llama 3 8B Instruct | Apr 17, 2024 | 8B | 8K |
Text input
Text output
|
★★★ | ★★ | $ |
Meta: Llama 3 70B Instruct | Apr 17, 2024 | 70B | 8K |
Text input
Text output
|
★★★★ | ★★ | $$$ |
Meta: Llama 2 70B Chat Unavailable | Jun 19, 2023 | 70B | 4K |
Text input
Text output
|
— | — | $$$$ |