Z AI
GLM-5.1
Compared with GLM-5, GLM-5.1 delivers significant improvements in coding, agentic tool usage, reasoning, role-play, and general chat quality. Besides, GLM-5.1 has outstanding capabilities in long-horizon agentic tasks like CUDA kernel optimization.
πVisionπ§ Reasoningπ§Tool callingβ‘Caching
Pricing per 1M tokens
Input
$1.40
Output
$4.40
Cache write
$4.40
Cache read
$0.26
Specifications
Context window200K tokens
Max output128K tokens
API typechat
AddedApr 7, 2026
Model IDzai/GLM-5.1
Privacy & data
Data retentionNo
Used for trainingNo
Provider locationπΈπ¬ Singapore
Privacy policyZ AI Privacy Policy β
