The Zhipu AI open platform has announced the free release of the GLM-4-Flash model, aimed at promoting innovation in AI applications. This model is known for its speed and cost-effectiveness, suitable for quick response to simple tasks, supporting multi-turn dialogue, web browsing, function calls, and long-text reasoning, covering 26 languages. Through optimization techniques such as adaptive weight quantization, parallelization, batching, and speculative sampling, GLM-4-Flash significantly reduces inference costs and has acquired over 10TB of high-quality multilingual data during the pre-training stage.