Z.ai: GLM 4.7

Text input Text output
Author's Description

GLM-4.7 is Z.ai’s latest flagship model, featuring upgrades in two key areas: enhanced programming capabilities and more stable multi-step reasoning/execution. It demonstrates significant improvements in executing complex agent tasks while delivering more natural conversational experiences and superior front-end aesthetics.

Key Specifications
Cost
$$$$$
Context
200K
Released
Dec 21, 2025
Speed
Ability
Reliability
Supported Parameters

This model supports the following parameters:

Response Format Tools Reasoning Top P Max Tokens Include Reasoning Tool Choice Temperature
Features

This model supports the following features:

Reasoning Tools Response Format
Performance Summary

Z.AI's GLM-4.7, released December 21, 2025, is positioned as a premium model with a context length of 200,000. It exhibits exceptional reliability, boasting a 98% success rate across benchmarks, indicating consistent operational stability. However, its speed is a notable weakness, with response times ranking in the 5th percentile. Similarly, its pricing is at the premium end, ranking in the 12th percentile. Despite its slower speed and higher cost, GLM-4.7 demonstrates strong performance in several key areas. It excels in Instruction Following (91st percentile accuracy), Reasoning (80th percentile accuracy), and Mathematics (82nd percentile accuracy), aligning with its description of enhanced programming capabilities and stable multi-step reasoning. Its perfect 100% accuracy in Ethics is particularly impressive, making it the most accurate model at its price point and among models of comparable speed. General Knowledge is also solid at 99.5% accuracy (70th percentile). A key area for improvement is its hallucination rate, with 88.0% accuracy placing it in the 36th percentile, suggesting it could better acknowledge uncertainty.

Model Pricing

Current Pricing

Feature Price (per 1M tokens)
Prompt $0.6
Completion $2.2
Input Cache Read $0.11

Price History

Available Endpoints
Provider Endpoint Name Context Length Pricing (Input) Pricing (Output)
Z.AI
Z.AI | z-ai/glm-4.7-20251222 200K $0.6 / 1M tokens $2.2 / 1M tokens
AtlasCloud
AtlasCloud | z-ai/glm-4.7-20251222 202K $0.52 / 1M tokens $1.75 / 1M tokens
Parasail
Parasail | z-ai/glm-4.7-20251222 202K $0.45 / 1M tokens $2.1 / 1M tokens
Chutes
Chutes | z-ai/glm-4.7-20251222 202K $0.4 / 1M tokens $1.5 / 1M tokens
Novita
Novita | z-ai/glm-4.7-20251222 204K $0.4 / 1M tokens $1.5 / 1M tokens
DeepInfra
DeepInfra | z-ai/glm-4.7-20251222 202K $0.4 / 1M tokens $1.75 / 1M tokens
GMICloud
GMICloud | z-ai/glm-4.7-20251222 202K $0.4 / 1M tokens $1.5 / 1M tokens
SiliconFlow
SiliconFlow | z-ai/glm-4.7-20251222 204K $0.42 / 1M tokens $2.2 / 1M tokens
Mancer 2
Mancer 2 | z-ai/glm-4.7-20251222 131K $0.4 / 1M tokens $1.5 / 1M tokens
Novita
Novita | z-ai/glm-4.7-20251222 204K $0.54 / 1M tokens $1.98 / 1M tokens
Cerebras
Cerebras | z-ai/glm-4.7-20251222 200K $0.4 / 1M tokens $1.5 / 1M tokens
Together
Together | z-ai/glm-4.7-20251222 202K $0.45 / 1M tokens $2 / 1M tokens
Venice
Venice | z-ai/glm-4.7-20251222 198K $0.55 / 1M tokens $2.65 / 1M tokens
Cerebras
Cerebras | z-ai/glm-4.7-20251222 131K $2.25 / 1M tokens $2.75 / 1M tokens
Google
Google | z-ai/glm-4.7-20251222 200K $0.6 / 1M tokens $2.2 / 1M tokens
Nebius
Nebius | z-ai/glm-4.7-20251222 202K $0.4 / 1M tokens $2 / 1M tokens
Mancer 2
Mancer 2 | z-ai/glm-4.7-20251222 131K $0.4 / 1M tokens $1.5 / 1M tokens
NextBit
NextBit | z-ai/glm-4.7-20251222 202K $0.6 / 1M tokens $2.5 / 1M tokens
Inceptron
Inceptron | z-ai/glm-4.7-20251222 202K $0.4 / 1M tokens $1.9 / 1M tokens
Benchmark Results
Benchmark Category Reasoning Strategy Free Executions Accuracy Cost Duration
Other Models by z-ai