Model Guide10 min readReviewed Apr 20, 2026

GLM-5 and GLM-5.1: Zhipu's Open-Source Flagship Models for Agentic Coding

GLM-5 launched on February 11, 2026 as Zhipu AI's open-source MoE flagship: 744B total parameters with 40B active, 28.5T training tokens, a 128K context window, and an MIT license. Less than two months later GLM-5.1 pushed further, reaching ~754B total / ~42B active parameters and demonstrating the ability to work autonomously for up to 8 hours on complex coding tasks. Together they form the current backbone of Zhipu's coding-agent ecosystem.

Published Apr 19, 2026Updated Apr 20, 2026
  • GLM-5 is a 744B / 40B active MoE model trained on 28.5T tokens under the MIT license.
  • GLM-5.1 can work autonomously for 8 hours and scores 58.4 on SWE-bench Pro, beating GPT-5.4.
  • GLM-5V-Turbo adds multimodal vision+coding with a 200K context window and 128K max output.
Quick note: This guide is based on public docs and release pages, but you should still verify current pricing, limits, supported tools, and region-specific billing on the official source before you pay, subscribe, or integrate.

Architecture and release timeline

GLM-5 was released on February 11, 2026 as Zhipu's largest open-source model to date. It uses a Mixture-of-Experts architecture with 744B total parameters and 40B active parameters per token, trained on 28.5 trillion tokens with a 128K context window. The model is released under the permissive MIT license, making it one of the most accessible frontier-scale open-source models available.

GLM-5.1 followed on April 8, 2026, scaling to approximately 754B total parameters and ~42B active parameters. The headline capability is sustained autonomous work: GLM-5.1 can operate independently on complex coding tasks for up to 8 hours. On SWE-bench Pro it scores 58.4, surpassing GPT-5.4 at 57.7.

GLM-5 family official snapshot infographic
A release-first view of what the public GLM docs actually confirm for GLM-5, GLM-5.1, and the DevPack route around them. Source: Official GLM-5 docs.
Official GLM-5.1 docs screenshot

Official screenshot

The GLM-5.1 docs are the cleanest public page for current flagship specs and pricing

The current GLM docs surface model context, feature support, and live API pricing more clearly than most third-party summaries. It is the safest place to anchor GLM family claims before branching into DevPack.

  • Useful for separating model docs from DevPack package messaging.
  • Helps keep GLM-5.1 and the wider GLM family aligned with official wording.

Source: Official GLM-5.1 docs.

GLM-5 and GLM-5.1 architecture comparison
ModelTotal ParamsActive ParamsContextRelease Date
GLM-5744B40B128KFeb 11, 2026
GLM-5.1~754B~42B128KApr 8, 2026
GLM-5V-TurboMultimodal200KApr 2, 2026

Benchmark performance

GLM-5 posts strong numbers across the main coding-agent benchmarks. On SWE-bench Verified it reaches 77.8, placing it competitively against Claude Opus 4.5 (80.9), GPT-5.2 (80.0), and MiniMax M2.5 (80.2). It does not lead every chart, but it consistently ranks in the top tier.

On agent-heavy benchmarks the picture is similarly competitive. Terminal-Bench 2.0 gives GLM-5 a score of 56.2 (60.7 on the extended evaluation), and BrowseComp with context comes in at 75.9, which is notably higher than Claude Opus 4.5 (67.8) and GPT-5.2 (65.8).

SWE-bench Verified

Public SWE-bench Verified scores for the current generation of coding-agent models.

Claude Opus 4.580.9

Official Anthropic evaluation.

MiniMax M2.580.2

Official MiniMax M2.5 release.

GPT-5.280.0

Official OpenAI evaluation.

GLM-577.8

Official Z.AI evaluation.

Kimi K2.576.8

Official Kimi K2.5 technical blog.

Source: Official Z.AI GLM-5 blog post.

Terminal-Bench 2.0

Multi-step terminal and agent workflow benchmark comparing leading coding models.

Claude Opus 4.559.3

Official Anthropic evaluation.

MiniMax M2.757.0

Official MiniMax M2.7 release.

GLM-556.2 / 60.7

Official Z.AI evaluation, standard / extended.

Kimi K2.550.8

Official Kimi K2.5 technical blog.

Source: Official Z.AI GLM-5 blog post.

BrowseComp w/ Context

Web browsing and information retrieval benchmark with context-augmented evaluation.

GLM-575.9

Official Z.AI evaluation.

Kimi K2.574.9

Official Kimi K2.5 technical blog.

Claude Opus 4.567.8

Official Anthropic evaluation.

GPT-5.265.8

Official OpenAI evaluation.

Source: Official Z.AI GLM-5 blog post.

Supported tools and integrations

GLM-5 and GLM-5.1 are designed to slot into existing coding-agent toolchains. Zhipu has documented support for a range of popular tools, and the DevPack subscription route provides a single entry point for accessing GLM models across all of them.

Tool support for GLM-5 / GLM-5.1
ToolCompatibilityNotes
Claude CodeAnthropic-compatibleUse the Anthropic route with model override
OpenCodeOpenAI-compatibleProvider picker or manual config
ClineOpenAI-compatibleStandard OpenAI base URL override
DroidOpenAI-compatibleSupported through coding endpoint
OpenClawDevPack routeFull agent workflow support
Kilo CodeOpenAI-compatibleSupported through coding endpoint
Roo CodeOpenAI-compatibleStandard OpenAI base URL override

API pricing

The public Z.AI pricing page already publishes USD token rows for the current GLM family, which makes it a cleaner buyer reference than older reposted CNY tables. The page also distinguishes direct API pricing from the separate DevPack package route.

  • DevPack subscriptions are still the package route for tool-first buying, while the pricing page above is the direct API route.
  • The official pricing page also keeps GLM-5.1, GLM-5, GLM-5-Turbo, and GLM-5V-Turbo on one family table.
  • For buyer-facing writing, explain the route first: DevPack for package access, pricing page for PAYG API billing.
Official GLM public API pricing (USD per 1M tokens)
ModelInputCached inputCache writeOutput
GLM-5.1$1.4$0.26Limited-time free$4.4
GLM-5$1.0$0.20Limited-time free$3.2
GLM-5V-Turbo$1.2$0.24Limited-time free$4.0
BuyGLM shows package prices in USD. When a source page is published in CNY, the displayed value uses a fixed 1 USD = 8 CNY conversion and should still be checked against the live vendor page before payment.

Start with DevPack to evaluate GLM-5.1 in your own workflow

The fastest way to test GLM-5.1 is through the DevPack subscription, which bundles API access with tool integration and avoids the need to configure endpoints manually.

Sources and official links

Frequently asked questions

What is the difference between GLM-5 and GLM-5.1?

GLM-5 is the base open-source model released Feb 11, 2026. GLM-5.1 is the improved version released Apr 8, 2026 with slightly more parameters (~754B total / ~42B active) and the ability to work autonomously for up to 8 hours. GLM-5.1 also scores higher on SWE-bench Pro at 58.4.

Is GLM-5 open source?

Yes. GLM-5 is released under the MIT license, making it one of the most permissively licensed frontier-scale models available. The weights are available on Hugging Face.

How does GLM-5 compare to Claude Opus 4.5 on SWE-bench?

GLM-5 scores 77.8 on SWE-bench Verified, while Claude Opus 4.5 scores 80.9. GLM-5 is competitive but does not lead on that specific benchmark. However, on BrowseComp with context, GLM-5 scores 75.9 vs Claude Opus 4.5 at 67.8, where GLM-5 does lead.