GLM-5V-Turbo is Z.AI's first multimodal coding model. It understands images, video, files, and UI layouts, then turns that visual context into runnable code, debugging help, and stronger agent workflows with Claude Code and OpenClaw.
GLM-5-Turbo is Z.ai’s high-speed variant of GLM-5, deeply optimized for OpenClaw from the training stage. It excels at precise tool calling, complex command following, scheduled and persistent tasks, and long-chain execution with near-zero hallucinations. Faster, more reliable, and purpose-built for real agent workflows.
A 744B MoE model (40B active) built for complex systems & agentic tasks. #1 open-source on Vending Bench 2, narrowing the gap with Claude Opus 4.5. Features DeepSeek Sparse Attention and "slime" RL infra.
GLM-4.5 is a new 355B parameter open-weight MoE model (32B active). It delivers state-of-the-art performance on reasoning, code, and agentic tasks. Both the 355B flagship and a 106B Air version are now available, featuring dual-mode inference.
GLM-4.6V is GLM's newest open-source multimodal model with a 128k context window. It features native function calling, bridging visual perception with executable actions for complex agentic workflows like web search and coding.
A lightweight (0.9B) professional OCR model. Achieves SOTA (94.6 on OmniDocBench) on complex layouts, tables, and handwriting. Supports vLLM/SGLang for ultra-fast inference.
GLM-Image combines a 9B Auto-regressive model with a 7B Diffusion decoder. This hybrid architecture excels at knowledge-dense generation, perfect for posters, diagrams, and precise text rendering. Open-source and ready for T2I & I2I tasks.
GLM-4.7 is a SOTA open-weight model optimized for coding and reasoning. It features "Preserved Thinking" to maintain reasoning context across multi-turn agentic tasks. Compatible with tools like Cline and Claude Code.
Official Zhipu AI platform (formerly ChatGLM team) to experience their new, MIT-licensed GLM models (Base, Reasoning, Rumination). Simple UI focuses on model interaction. Free.