Tongyi DeepResearch 30B A3B

Text input Text output
Author's Description

Tongyi DeepResearch is an agentic large language model developed by Tongyi Lab, with 30 billion total parameters activating only 3 billion per token. It's optimized for long-horizon, deep information-seeking tasks and delivers state-of-the-art performance on benchmarks like Humanity's Last Exam, BrowserComp, BrowserComp-ZH, WebWalkerQA, GAIA, xbench-DeepSearch, and FRAMES. This makes it superior for complex agentic search, reasoning, and multi-step problem-solving compared to prior models. The model includes a fully automated synthetic data pipeline for scalable pre-training, fine-tuning, and reinforcement learning. It uses large-scale continual pre-training on diverse agentic data to boost reasoning and stay fresh. It also features end-to-end on-policy RL with a customized Group Relative Policy Optimization, including token-level gradients and negative sample filtering for stable training. The model supports ReAct for core ability checks and an IterResearch-based 'Heavy' mode for max performance through test-time scaling. It's ideal for advanced research agents, tool use, and heavy inference workflows.

Key Specifications
Context
131K
Parameters
30B
Released
Sep 18, 2025
Supported Parameters

This model supports the following parameters:

Response Format Structured Outputs Top P Tools Temperature Reasoning Tool Choice Include Reasoning Max Tokens
Features

This model supports the following features:

Tools Response Format Reasoning Structured Outputs
Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.09
Completion $0.45

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
AtlasCloud
AtlasCloud | alibaba/tongyi-deepresearch-30b-a3b 131K $0.09 / 1M tokens $0.45 / 1M tokens