PyTorch implementation for math problem-solving LLM research
Top 85.5% on sourcepulse
This repository provides the official PyTorch implementation for MathGLM, a family of large language models designed to excel at mathematical tasks, including arithmetic operations and word problems. It challenges the notion that LLMs struggle with calculations, demonstrating high accuracy on multi-digit arithmetic and competitive performance on math word problems, targeting researchers and developers working with LLMs for quantitative reasoning.
How It Works
MathGLM models are fine-tuned from existing GLM architectures (e.g., GLM-10B, ChatGLM) on specialized datasets containing multi-step arithmetic operations and text-based math problems. This approach aims to imbue the models with robust quantitative reasoning capabilities, enabling them to perform complex calculations and solve word problems without external tools, thereby surpassing the arithmetic accuracy of models like GPT-4 in specific benchmarks.
Quick Start & Requirements
conda env create -f env.yml
.deepspeed
(v0.6.0 for arithmetic tasks, v0.9.5 for 6B MWP tasks), and SwissArmyTransformer
../inference.sh
within MathGLM_Arithmetic
or MathGLM_MWP
directories.Highlighted Details
Maintenance & Community
The project is associated with THUDM (Tsinghua University Knowledge Engineering Group). Further community engagement details are not explicitly provided in the README.
Licensing & Compatibility
The repository does not explicitly state a license. The models are available via THU-Cloud and ModelScope. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The README notes specific deepspeed
version requirements for different tasks, indicating potential dependency sensitivity. Performance on English math word problems is not detailed.
1 year ago
1+ week