Academic headlines report that researchers from Tsinghua University, TAL AI Lab, and Zhipu AI have proposed a 2-billion-parameter language model called MathGLM, aimed at exploring the efficiency of large language models in mathematical reasoning. The model employs a Transformer decoder architecture and has been trained on a large-scale arithmetic dataset, significantly enhancing its mathematical operation capabilities. Experimental results show that MathGLM achieves near 100% accuracy on a series of arithmetic tasks, outperforming GPT-4. Even with only 100 million parameters, MathGLM surpasses both GPT-4 and ChatGPT. The study also found that as the number of parameters increases, MathGLM's arithmetic operation abilities also improve. When handling complex mixed arithmetic operations with intricate number formats, MathGLM also outperforms GPT-4 and ChatGPT. This research indicates that under the condition of sufficient parameters and data volume, language models can accurately perform complex mathematical operations.
Stronger than GPT-4, 2 Billion Parameter Model Achieves Nearly 100% Accuracy in Arithmetic Tasks

学术头条
This article is from AIbase Daily
Welcome to the [AI Daily] column! This is your daily guide to exploring the world of artificial intelligence. Every day, we present you with hot topics in the AI field, focusing on developers, helping you understand technical trends, and learning about innovative AI product applications.