Perplexity: Llama 3.1 Sonar 70B Online

Text input Text output Unavailable
Author's Description

Llama 3.1 Sonar is Perplexity's latest model family. It surpasses their earlier Sonar models in cost-efficiency, speed, and performance. This is the online version of the [offline chat model](/models/perplexity/llama-3.1-sonar-large-128k-chat). It is focused on delivering helpful, up-to-date, and factual responses. #online

Key Specifications
Cost
$$$$
Context
127K
Parameters
70B
Released
Jul 31, 2024
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Top P Frequency Penalty Max Tokens Presence Penalty Temperature
Performance Summary

Perplexity: Llama 3.1 Sonar 70B Online, released on July 31, 2024, demonstrates moderate performance in both speed and pricing, ranking in the 33rd and 39th percentiles respectively across five benchmarks. This online model, designed for factual and up-to-date responses, shows a mixed performance across various categories. It excels in Email Classification with 97.0% accuracy, indicating strong capabilities in understanding context and categorizing information. However, its performance in Coding (1.0% accuracy) and General Knowledge (8.5% accuracy) is notably low, suggesting significant weaknesses in these domains. Reasoning and Ethics benchmarks show moderate accuracy at 48.0% and 81.0% respectively, with the Ethics performance being somewhat below average for its percentile. The model's cost and duration metrics are generally moderate across benchmarks, aligning with its overall pricing and speed rankings. Its primary strength lies in classification tasks, while its main weaknesses are in complex coding and broad general knowledge recall.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $1
Completion $1
Request $0.005 per request

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
Perplexity
Perplexity | perplexity/llama-3.1-sonar-large-128k-online 127K $1 / 1M tokens $1 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Free Executions Accuracy Cost Duration
Other Models by perplexity