Meta: Llama 3.2 1B Instruct

Text input Text output
Author's Description

Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate...

Key Specifications
Cost
$
Context
131K
Parameters
1B
Released
Sep 24, 2024
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Seed Frequency Penalty Top P Min P Response Format Temperature Stop Presence Penalty Max Tokens
Features

This model supports the following features:

Response Format
Performance Summary

Meta's Llama 3.2 1B Instruct model, released on September 24, 2024, is a 1-billion-parameter language model designed for efficient natural language tasks in low-resource environments. It consistently ranks among the fastest models, achieving an Infinityth percentile across seven benchmarks, and offers highly competitive pricing, also at the Infinityth percentile. This makes it an exceptionally cost-effective and rapid solution. However, the model's performance on knowledge-based and complex reasoning tasks is a significant weakness, with 0.0% accuracy in General Knowledge, Ethics, and Coding, and only 22.0% in Reasoning. Its accuracy in Mathematics is also 0.0%. Strengths appear in more constrained tasks: Email Classification achieved 32.0% accuracy (5th percentile) with excellent cost and duration percentiles (96th and 93rd respectively). Instruction Following, while having a lower accuracy of 18.9% (21st percentile), also demonstrated strong cost efficiency (93rd percentile). Overall, Llama 3.2 1B is best suited for applications prioritizing speed and cost-efficiency in specific, less cognitively demanding NLP tasks, rather than complex problem-solving or knowledge retrieval.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.027
Completion $0.2

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
DeepInfra
DeepInfra | meta-llama/llama-3.2-1b-instruct 131K $0.027 / 1M tokens $0.2 / 1M tokens
InferenceNet
InferenceNet | meta-llama/llama-3.2-1b-instruct 16K $0.027 / 1M tokens $0.2 / 1M tokens
Cloudflare
Cloudflare | meta-llama/llama-3.2-1b-instruct 60K $0.027 / 1M tokens $0.2 / 1M tokens
SambaNova
SambaNova | meta-llama/llama-3.2-1b-instruct 16K $0.027 / 1M tokens $0.2 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Strategy Free Executions Accuracy Cost Duration
Other Models by meta-llama