Discover and explore top open-source AI tools and projects—updated daily.
TIGER-AI-LabLLM for math problem-solving, targeting generalizability
Top 75.3% on SourcePulse
MAmmoTH provides open-source large language models (LLMs) specialized for mathematical reasoning, built upon instruction tuning with a novel hybrid Chain-of-Thought (CoT) and Program-of-Thought (PoT) approach. It targets researchers and developers aiming to enhance LLM performance on diverse mathematical tasks, offering models based on Llama-2, Code Llama, and Mistral architectures.
How It Works
MAmmoTH models are trained on the MathInstruct dataset, which emphasizes a hybrid CoT/PoT rationale strategy. This approach allows the models to generate executable code (PoT) for problem-solving when feasible, falling back to CoT reasoning otherwise. This hybrid decoding method aims to improve accuracy and robustness across a wide range of mathematical problems.
Quick Start & Requirements
pip install -r requirements.txttransformers, datasets, vllm (for optimized inference). GPU is recommended for training and efficient inference.Highlighted Details
Maintenance & Community
The project is associated with TIGER-AI-Lab. Further community engagement details are not explicitly provided in the README.
Licensing & Compatibility
The project's dataset licenses vary, including MIT, Apache 2.0, and non-commercial licenses for some subsets. Commercial use may be restricted depending on the specific dataset components utilized.
Limitations & Caveats
The README notes that some dataset subsets have non-listed or non-commercial licenses, requiring careful review for commercial applications. Performance can vary significantly based on the chosen base model and decoding strategy.
1 year ago
1 day
PhoebusSi
allenai
nlpxucan
deepseek-ai
deepseek-ai