Meta: Llama 3.2 1B Instruct

Text input Text output
Author's Description

Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate efficiently in low-resource environments while maintaining strong task performance. Supporting eight core languages and fine-tunable for more, Llama 1.3B is ideal for businesses or developers seeking lightweight yet powerful AI solutions that can operate in diverse multilingual settings without the high computational demand of larger models. Click here for the [original model card](https://github.com/meta-llama/llama-models/blob/main/models/llama3_2/MODEL_CARD.md). Usage of this model is subject to [Meta's Acceptable Use Policy](https://www.llama.com/llama3/use-policy/).

Key Specifications
Cost
$
Context
131K
Parameters
1B
Released
Sep 24, 2024
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Stop Presence Penalty Top P Temperature Seed Min P Response Format Frequency Penalty Max Tokens
Features

This model supports the following features:

Response Format
Performance Summary

Meta's Llama 3.2 1B Instruct model consistently ranks among the fastest models and offers highly competitive pricing across all evaluated benchmarks. This 1-billion-parameter model, designed for efficiency in low-resource environments, demonstrates its cost-effectiveness particularly well. While excelling in cost and speed, its accuracy performance varies significantly across categories. It achieved a notable 32.0% accuracy in Email Classification, placing it in the 6th percentile for accuracy but impressively within the top 3 for cost and 93rd percentile for duration, indicating strong efficiency for this task. Similarly, in Reasoning, it showed a 28.6% accuracy (16th percentile) but again secured a top 3 position for cost and the best accuracy-to-cost ratio. However, the model exhibited very low accuracy in Instruction Following (18.9%), and zero accuracy in Coding, Ethics, and General Knowledge, suggesting significant limitations in these complex domains. Its strengths lie in its operational efficiency and cost-effectiveness, making it suitable for tasks where these factors are paramount, especially in classification and basic reasoning, but less so for tasks requiring deep knowledge or complex problem-solving.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.005
Completion $0.01

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
DeepInfra
DeepInfra | meta-llama/llama-3.2-1b-instruct 131K $0.005 / 1M tokens $0.01 / 1M tokens
InferenceNet
InferenceNet | meta-llama/llama-3.2-1b-instruct 16K $0.01 / 1M tokens $0.01 / 1M tokens
Cloudflare
Cloudflare | meta-llama/llama-3.2-1b-instruct 60K $0.027 / 1M tokens $0.2 / 1M tokens
SambaNova
SambaNova | meta-llama/llama-3.2-1b-instruct 16K $0.005 / 1M tokens $0.01 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Free Executions Accuracy Cost Duration
Other Models by meta-llama