SOTA LLM for math problem solving
Top 83.6% on sourcepulse
Abel is an open-source Large Language Model (LLM) focused on achieving state-of-the-art performance in mathematical reasoning without relying on external tools, reward models, or RLHF. It targets researchers and developers working on AI for STEM education and complex problem-solving, offering significant improvements over existing models on benchmarks like GSM8K and MATH.
How It Works
Abel is trained using a novel Supervised Fine-Tuning (SFT) methodology called "Parental Oversight." This approach emphasizes data processing philosophy, treating fine-tuning data like educational methods for children. It prioritizes data quality, relevance, and the inclusion of step-by-step reasoning, aiming to instill a deeper understanding rather than just memorization. This SFT-centric approach is presented as a significantly underestimated method for achieving high performance in complex reasoning tasks.
Quick Start & Requirements
conda create -n abel python=3.10
), activate it (conda activate abel
), and install dependencies (pip install -r requirements.txt
).bash evaluation/eval.sh
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The model's generalization capabilities are limited to specific mathematical domains, lacking broad applicability to diverse problem types or integration into multi-domain chatbots. Multilingual support is absent, and advanced techniques like reward models and RLHF have not been explored.
1 year ago
1 day