Author's Description
Compared with GLM-4.5, this generation brings several key improvements: Longer context window: The context window has been expanded from 128K to 200K tokens, enabling the model to handle more complex agentic tasks. Superior coding performance: The model achieves higher scores on code benchmarks and demonstrates better real-world performance in applications such as Claude Code、Cline、Roo Code and Kilo Code, including improvements in generating visually polished front-end pages. Advanced reasoning: GLM-4.6 shows a clear improvement in reasoning performance and supports tool use during inference, leading to stronger overall capability. More capable agents: GLM-4.6 exhibits stronger performance in tool using and search-based agents, and integrates more effectively within agent frameworks. Refined writing: Better aligns with human preferences in style and readability, and performs more naturally in role-playing scenarios.
Key Specifications
Supported Parameters
This model supports the following parameters:
Features
This model supports the following features:
Performance Summary
Z.AI: GLM 4.6, released by z-ai on September 30, 2025, is a significant upgrade from GLM-4.5, featuring an expanded 200K token context window for complex agentic tasks. This model demonstrates exceptional reliability with a 97% success rate across benchmarks, indicating consistent operational stability. However, it tends to have longer response times, ranking in the 10th percentile for speed, and is positioned at premium pricing levels (12th percentile). Performance highlights include perfect accuracy in General Knowledge and Email Classification, with these benchmarks also noted for being the most accurate at their respective price points and speeds. GLM 4.6 excels in Coding (95.0% accuracy, 94th percentile) and Mathematics (96.0% accuracy, 98th percentile), aligning with its described superior coding performance and advanced reasoning capabilities. Its hallucination rate is low at 96.0% accuracy. A notable weakness is its Instruction Following, where it scored only 3.1% accuracy (22nd percentile), suggesting challenges with multi-layered or complex directives. Ethics performance is also relatively low at 82.0% (21st percentile).
Model Pricing
Current Pricing
| Feature | Price (per 1M tokens) |
|---|---|
| Prompt | $0.6 |
| Completion | $2.2 |
| Input Cache Read | $0.11 |
Price History
Available Endpoints
| Provider | Endpoint Name | Context Length | Pricing (Input) | Pricing (Output) |
|---|---|---|---|---|
|
Z.AI
|
Z.AI | z-ai/glm-4.6 | 200K | $0.6 / 1M tokens | $2.2 / 1M tokens |
|
Parasail
|
Parasail | z-ai/glm-4.6 | 202K | $0.35 / 1M tokens | $1.5 / 1M tokens |
|
DeepInfra
|
DeepInfra | z-ai/glm-4.6 | 202K | $0.35 / 1M tokens | $1.5 / 1M tokens |
|
Chutes
|
Chutes | z-ai/glm-4.6 | 202K | $0.35 / 1M tokens | $1.5 / 1M tokens |
|
GMICloud
|
GMICloud | z-ai/glm-4.6 | 204K | $0.35 / 1M tokens | $1.5 / 1M tokens |
|
Novita
|
Novita | z-ai/glm-4.6 | 204K | $0.35 / 1M tokens | $1.5 / 1M tokens |
|
SiliconFlow
|
SiliconFlow | z-ai/glm-4.6 | 204K | $0.39 / 1M tokens | $1.9 / 1M tokens |
|
AtlasCloud
|
AtlasCloud | z-ai/glm-4.6 | 202K | $0.44 / 1M tokens | $1.74 / 1M tokens |
|
Mancer 2
|
Mancer 2 | z-ai/glm-4.6 | 131K | $0.45 / 1M tokens | $2 / 1M tokens |
|
Novita
|
Novita | z-ai/glm-4.6 | 204K | $0.55 / 1M tokens | $2.2 / 1M tokens |
|
BaseTen
|
BaseTen | z-ai/glm-4.6 | 200K | $0.6 / 1M tokens | $2.2 / 1M tokens |
|
Fireworks
|
Fireworks | z-ai/glm-4.6 | 202K | $0.55 / 1M tokens | $2.19 / 1M tokens |
|
Chutes
|
Chutes | z-ai/glm-4.6 | 202K | $0.35 / 1M tokens | $1.5 / 1M tokens |
|
DeepInfra
|
DeepInfra | z-ai/glm-4.6 | 202K | $0.43 / 1M tokens | $1.75 / 1M tokens |
|
Friendli
|
Friendli | z-ai/glm-4.6 | 131K | $0.6 / 1M tokens | $2.2 / 1M tokens |
|
Cerebras
|
Cerebras | z-ai/glm-4.6 | 131K | $2.25 / 1M tokens | $2.75 / 1M tokens |
|
Together
|
Together | z-ai/glm-4.6 | 202K | $0.6 / 1M tokens | $2.2 / 1M tokens |
|
Avian
|
Avian | z-ai/glm-4.6 | 204K | $0.35 / 1M tokens | $1.5 / 1M tokens |
|
Avian
|
Avian | z-ai/glm-4.6 | 202K | $0.35 / 1M tokens | $1.5 / 1M tokens |
Benchmark Results
| Benchmark | Category | Reasoning | Strategy | Free | Executions | Accuracy | Cost | Duration |
|---|
Other Models by z-ai
|
|
Released | Params | Context |
|
Speed | Ability | Cost |
|---|---|---|---|---|---|---|---|
| Z.AI: GLM 4.7 | Dec 21, 2025 | — | 200K |
Text input
Text output
|
★ | ★★★★★ | $$$$$ |
| Z.AI: GLM 4.6V | Dec 08, 2025 | — | 131K |
Image input
Video input
Text input
Text output
|
★★ | ★★★★★ | $$$$ |
| Z.AI: GLM 4.6 (exacto) | Sep 30, 2025 | — | 202K |
Text input
Text output
|
— | — | $$$$ |
| Z.AI: GLM 4.5V | Aug 11, 2025 | ~106B | 65K |
Image input
Text input
Text output
|
★★ | ★★★ | $$$$$ |
| Z.AI: GLM 4.5 | Jul 25, 2025 | — | 131K |
Text input
Text output
|
★ | ★★★★ | $$$$$ |
| Z.AI: GLM 4.5 Air | Jul 25, 2025 | — | 131K |
Text input
Text output
|
★ | ★★ | $$$$$ |
| Z.AI: GLM 4 32B | Jul 24, 2025 | 32B | 128K |
Text input
Text output
|
★★★ | ★ | $$ |