AI21: Jamba Mini 1.6

Text input Text output Unavailable
Author's Description

AI21 Jamba Mini 1.6 is a hybrid foundation model combining State Space Models (Mamba) with Transformer attention mechanisms. With 12 billion active parameters (52 billion total), this model excels in extremely long-context tasks (up to 256K tokens) and achieves superior inference efficiency, outperforming comparable open models on tasks such as retrieval-augmented generation (RAG) and grounded question answering. Jamba Mini 1.6 supports multilingual tasks across English, Spanish, French, Portuguese, Italian, Dutch, German, Arabic, and Hebrew, along with structured JSON output and tool-use capabilities. Usage of this model is subject to the [Jamba Open Model License](https://www.ai21.com/licenses/jamba-open-model-license).

Key Specifications
Cost
$$
Context
256K
Parameters
52B (Rumoured)
Released
Mar 13, 2025
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Tools Temperature Max Tokens Top P Tool Choice Stop
Features

This model supports the following features:

Tools
Performance Summary

AI21 Jamba Mini 1.6 demonstrates strong performance in terms of efficiency, consistently ranking among the fastest models with a 73rd percentile speed ranking across five benchmarks. It also offers competitive pricing, placing in the 70th percentile for cost-effectiveness. The model's hybrid architecture, combining State Space Models and Transformer attention, is designed for extremely long-context tasks, supporting up to 256K tokens, and boasts superior inference efficiency. However, its accuracy across various benchmarks is a notable weakness. In General Knowledge, it achieved only 2.0% accuracy (10th percentile), and in Ethics, 34.0% (13th percentile). Coding performance was also low at 7.0% accuracy (14th percentile). While Email Classification showed a respectable 90.0% accuracy, this still placed it in the 17th percentile. Instruction Following was moderate at 41.0% accuracy (40th percentile). Despite these accuracy challenges, its efficiency in terms of duration and cost for these tasks remains competitive. Key strengths lie in its architectural innovation for long-context processing, multilingual support, structured JSON output, and tool-use capabilities, making it potentially valuable for specific applications where these features are paramount, even with current accuracy limitations.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.2
Completion $0.4

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
AI21
AI21 | ai21/jamba-1.6-mini 256K $0.2 / 1M tokens $0.4 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Strategy Free Executions Accuracy Cost Duration
Other Models by ai21