Math LLM for bilingual reasoning tasks
Top 61.5% on sourcepulse
InternLM-Math provides state-of-the-art bilingual Large Language Models (LLMs) specifically designed for mathematical reasoning. It functions as a solver, prover, verifier, and augmentor, targeting researchers and developers working on advanced mathematical AI capabilities. The models offer significant improvements in both informal and formal mathematical reasoning tasks.
How It Works
InternLM-Math models are continued pre-trained on approximately 100 billion math-related tokens and then fine-tuned on a large bilingual math dataset. A key innovation is the integration of Lean 4, a theorem proving language, enabling the models to generate Lean code for reasoning tasks and act as a reward model for verifying reasoning steps. This approach allows for verifiable mathematical reasoning and auto-formalization.
Quick Start & Requirements
transformers
library is also supported.Highlighted Details
Maintenance & Community
The project is actively developed by Shanghai AI Laboratory and associated researchers. Recent updates include new models like InternLM2.5-Step-Prover and datasets like Lean-Workbook.
Licensing & Compatibility
The project is released under the Apache 2.0 license, permitting commercial use and modification.
Limitations & Caveats
The models may exhibit issues such as jumping calculation steps, performing poorly on specific Chinese fill-in-the-blank and English multiple-choice problems, and code switching due to SFT data composition. Lean-related abilities might be limited to specific problem types.
9 months ago
1 day