Chinese AI developer z.ai has launched GLM-5, a model that redefines the boundaries of reliability in large language models. Unlike predecessors focused solely on conversational accuracy, GLM-5 prioritizes actionable output—transforming prompts into fully formatted documents (.docx, .pdf, .xlsx) while maintaining an industry-leading hallucination rate of -1 on the AA-Omniscience Index, a 35-point improvement over GLM-4.5.

This shift isn’t just about smarter answers; it’s about autonomous execution. GLM-5 operates in Agent Mode, decomposing tasks into subtasks and generating professional-grade deliverables without manual intervention. A financial report, a sponsorship proposal, or a complex spreadsheet can now be produced directly from a prompt—bridging the gap between AI assistance and full-scale automation.

Behind the scenes, GLM-5’s architecture reflects its ambition. The model scales to 744 billion parameters (up from 355B in GLM-4.5), with a Mixture-of-Experts (MoE) design activating 40 billion parameters per token to optimize performance. Training leverages Slime, z.ai’s asynchronous reinforcement learning framework, which accelerates iteration cycles by breaking traditional RL bottlenecks. The result? A model that not only reasons better but executes at scale—a critical advancement for enterprise workflows.

  • SWE-bench Verified: 77.8 (outperforming Gemini 3 Pro’s 76.2, approaching Claude Opus 4.6’s 80.9).
  • Vending Bench 2: Simulated business management with a final balance of $4,432.12, topping open-source competitors.
  • Pricing: $0.80–$1.00 per million input tokens and $2.56–$3.20 per million output tokens—roughly 6x cheaper than Claude Opus 4.6 ($5/$25) and competitive with mid-tier models like Grok 4.1.

The cost efficiency extends beyond raw token pricing. GLM-5 integrates DeepSeek Sparse Attention (DSA), maintaining a 200K context window while reducing computational overhead—a necessity for models of this scale. For enterprises, this means deploying frontier AI without the prohibitive infrastructure costs of proprietary alternatives.

Who stands to benefit?

GLM-5 is designed for organizations that treat AI as a force multiplier, not just a tool. Engineers building self-healing pipelines, legal teams drafting contracts, or finance departments generating reports will find its document-generation capabilities transformative. The MIT License ensures no vendor lock-in, while the open-weights model allows for on-premise deployment—a strategic advantage in regulated industries.

Yet challenges remain. The model’s 744B parameters demand significant hardware resources, and its aggressive execution style—while efficient—lacks situational awareness in complex scenarios. Early feedback highlights a risk of over-optimization without context, a tradeoff that may require human oversight for critical applications.

For enterprises ready to move beyond chatbots and into autonomous workflows, GLM-5 offers a compelling proposition: state-of-the-art performance at a fraction of the cost, with the flexibility to integrate into existing systems. Whether it’s the future of office automation or a stepping stone toward AGI, one thing is clear—z.ai has just raised the bar for what AI can do.