A 744B MoE model (40B active) built for complex systems & agentic tasks. #1 open-source on Vending Bench 2, narrowing the gap with Claude Opus 4.5. Features DeepSeek Sparse Attention and "slime" RL infra.
A lightweight (0.9B) professional OCR model. Achieves SOTA (94.6 on OmniDocBench) on complex layouts, tables, and handwriting. Supports vLLM/SGLang for ultra-fast inference.
GLM-4.6V is GLM's newest open-source multimodal model with a 128k context window. It features native function calling, bridging visual perception with executable actions for complex agentic workflows like web search and coding.
GLM-4.5 is a new 355B parameter open-weight MoE model (32B active). It delivers state-of-the-art performance on reasoning, code, and agentic tasks. Both the 355B flagship and a 106B Air version are now available, featuring dual-mode inference.
GLM-Image combines a 9B Auto-regressive model with a 7B Diffusion decoder. This hybrid architecture excels at knowledge-dense generation, perfect for posters, diagrams, and precise text rendering. Open-source and ready for T2I & I2I tasks.
GLM-4.7 is a SOTA open-weight model optimized for coding and reasoning. It features "Preserved Thinking" to maintain reasoning context across multi-turn agentic tasks. Compatible with tools like Cline and Claude Code.
Official Zhipu AI platform (formerly ChatGLM team) to experience their new, MIT-licensed GLM models (Base, Reasoning, Rumination). Simple UI focuses on model interaction. Free.