Meta: Llama 3.1 70B Instruct

Text input Text output
Author's Description

Meta's latest class of model (Llama 3.1) launched with a variety of sizes & flavors. This 70B instruct-tuned version is optimized for high quality dialogue usecases. It has demonstrated strong performance compared to leading closed-source models in human evaluations. To read more about the model release, [click here](https://ai.meta.com/blog/meta-llama-3-1/). Usage of this model is subject to [Meta's Acceptable Use Policy](https://llama.meta.com/llama3/use-policy/).

Key Specifications
Cost
$$
Context
131K
Parameters
70B
Released
Jul 22, 2024
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Stop Presence Penalty Tool Choice Top P Temperature Seed Min P Tools Response Format Frequency Penalty Max Tokens
Features

This model supports the following features:

Tools Response Format
Performance Summary

Meta's Llama 3.1 70B Instruct model demonstrates a strong overall performance profile, particularly excelling in cost-efficiency and reliability. It consistently offers among the most competitive pricing, ranking in the 81st percentile across benchmarks, and exhibits high reliability with an 84% success rate, indicating consistent and usable responses. The model typically performs in the top tier for speed, ranking in the 61st percentile. Analysis of benchmark results reveals key strengths in specific areas. The model achieved perfect accuracy in the Ethics (Baseline) benchmark, making it the most accurate model at its price point and speed. It also performed exceptionally well in Email Classification (98.0% accuracy) and General Knowledge (97.8% accuracy). While strong in instruction following (54.0% accuracy), a notable weakness is observed in Coding (Baseline), where it achieved only 2.0% accuracy, placing it in the 10th percentile. Reasoning (Baseline) also shows room for improvement at 56.0% accuracy. This model is optimized for high-quality dialogue use cases, and its performance reflects this focus, particularly in areas requiring nuanced understanding and ethical considerations.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.1
Completion $0.28

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
DeepInfra
DeepInfra | meta-llama/llama-3.1-70b-instruct 131K $0.1 / 1M tokens $0.28 / 1M tokens
Lambda
Lambda | meta-llama/llama-3.1-70b-instruct 131K $0.12 / 1M tokens $0.3 / 1M tokens
Nebius
Nebius | meta-llama/llama-3.1-70b-instruct 131K $0.13 / 1M tokens $0.4 / 1M tokens
DeepInfra
DeepInfra | meta-llama/llama-3.1-70b-instruct 131K $0.23 / 1M tokens $0.4 / 1M tokens
InferenceNet
InferenceNet | meta-llama/llama-3.1-70b-instruct 16K $0.1 / 1M tokens $0.28 / 1M tokens
Hyperbolic
Hyperbolic | meta-llama/llama-3.1-70b-instruct 131K $0.4 / 1M tokens $0.4 / 1M tokens
Together
Together | meta-llama/llama-3.1-70b-instruct 131K $0.88 / 1M tokens $0.88 / 1M tokens
Fireworks
Fireworks | meta-llama/llama-3.1-70b-instruct 131K $0.9 / 1M tokens $0.9 / 1M tokens
NextBit
NextBit | meta-llama/llama-3.1-70b-instruct 131K $0.1 / 1M tokens $0.28 / 1M tokens
Phala
Phala | meta-llama/llama-3.1-70b-instruct 131K $0.1 / 1M tokens $0.28 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Free Executions Accuracy Cost Duration
Other Models by meta-llama