GLM-5 is a next-generation open-source large language model (LLM) developed by the Z .ai team under the zai-org organization that pushes the boundaries of reasoning, coding, and long-horizon agentic intelligence. Building on earlier GLM series models, GLM-5 dramatically scales the parameter count (to roughly 744 billion) and expands pre-training data to significantly improve performance on complex tasks such as multi-step reasoning, software engineering workflows, and agent orchestration compared to its predecessors like GLM-4.5. It incorporates innovations like DeepSeek Sparse Attention (DSA) to preserve massive context windows while reducing deployment costs and supporting long context processing, which is crucial for detailed plans and agent tasks.
Features
- Massive 744 billion-parameter pre-trained foundation model
- DeepSeek Sparse Attention for efficient long-context handling
- Enhanced reasoning and coding capabilities
- Improved agentic task performance and multi-step planning
- Support for efficient deployment via FP8 and vLLM
- Open-source, community-driven research model