Math reasoning model for competition-level problems
Top 17.2% on sourcepulse
DeepSeekMath provides open-source 7B parameter language models specifically trained for advanced mathematical reasoning. Targeting researchers and developers, these models offer strong performance on benchmarks like MATH, approaching proprietary model capabilities without external toolkits, and also demonstrate robust tool-use and coding abilities.
How It Works
DeepSeekMath models are initialized from DeepSeek-Coder-v1.5 7B and further pre-trained on a massive 500B token dataset comprising mathematical web texts, natural language, and code. This extensive training, particularly on curated mathematical content from Common Crawl, imbues the models with superior mathematical reasoning capabilities. The instruct and RL variants are fine-tuned for better instruction following and optimized using a Group Relative Policy Optimization (GRPO) algorithm, respectively.
Quick Start & Requirements
torch_dtype=torch.bfloat16
is recommended for inference.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
1 day