Meta: Llama 3.2 3B Instruct

Text input Text output Free Option
Author's Description

Llama 3.2 3B is a 3-billion-parameter multilingual large language model, optimized for advanced natural language processing tasks like dialogue generation, reasoning, and summarization. Designed with the latest transformer architecture, it supports eight languages, including English, Spanish, and Hindi, and is adaptable for additional languages. Trained on 9 trillion tokens, the Llama 3.2 3B model excels in instruction-following, complex reasoning, and tool use. Its balanced performance makes it ideal for applications needing accuracy and efficiency in text generation across multilingual settings. Click here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md). Usage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).

Key Specifications
Cost
$
Context
131K
Parameters
3B
Released
Sep 24, 2024
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Stop Presence Penalty Top P Temperature Seed Min P Response Format Frequency Penalty Max Tokens
Features

This model supports the following features:

Response Format
Performance Summary

Meta's Llama 3.2 3B Instruct, created on September 24, 2024, demonstrates strong performance in terms of efficiency and cost-effectiveness. It consistently ranks among the fastest models and offers highly competitive pricing, placing in the 99th percentile across multiple benchmarks. While its reliability is not explicitly detailed with a percentile, the provided benchmark results indicate consistent response generation. Across benchmark categories, Llama 3.2 3B shows varied performance. It excels in cost efficiency, frequently appearing in the top 3 for cost and achieving the best accuracy-to-cost ratio in General Knowledge. Its accuracy in Email Classification is notable at 88.0%, making it the most accurate model at its price point for this task. However, the model exhibits weaknesses in more complex cognitive tasks. Its accuracy is low in Coding (11.0%), Instruction Following (0.0% and 26.3%), Reasoning (27.3%), and Ethics (44.0%). Despite these lower accuracy scores, its low cost per query across all benchmarks remains a significant advantage, particularly for applications where cost-efficiency is paramount and a degree of error can be tolerated or mitigated.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.012
Completion $0.024

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
DeepInfra
DeepInfra | meta-llama/llama-3.2-3b-instruct 131K $0.012 / 1M tokens $0.024 / 1M tokens
Lambda
Lambda | meta-llama/llama-3.2-3b-instruct 131K $0.015 / 1M tokens $0.025 / 1M tokens
InferenceNet
InferenceNet | meta-llama/llama-3.2-3b-instruct 16K $0.02 / 1M tokens $0.02 / 1M tokens
Novita
Novita | meta-llama/llama-3.2-3b-instruct 32K $0.03 / 1M tokens $0.05 / 1M tokens
Cloudflare
Cloudflare | meta-llama/llama-3.2-3b-instruct 128K $0.051 / 1M tokens $0.34 / 1M tokens
Together
Together | meta-llama/llama-3.2-3b-instruct 131K $0.06 / 1M tokens $0.06 / 1M tokens
SambaNova
SambaNova | meta-llama/llama-3.2-3b-instruct 4K $0.003 / 1M tokens $0.006 / 1M tokens
Hyperbolic
Hyperbolic | meta-llama/llama-3.2-3b-instruct 131K $0.1 / 1M tokens $0.1 / 1M tokens
Nineteen
Nineteen | meta-llama/llama-3.2-3b-instruct 20K $0.003 / 1M tokens $0.006 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Free Executions Accuracy Cost Duration
Other Models by meta-llama