Discover and explore top open-source AI tools and projects—updated daily.
Gen-VerseLLM post-training algorithms for data selection, RL, and inference
Top 62.4% on SourcePulse
ReasonFlux introduces a novel template-augmented reasoning paradigm for Large Language Models (LLMs), aiming to enhance performance on complex reasoning tasks. It targets researchers and developers seeking to improve LLM capabilities in areas like mathematics and general question answering, offering a method to scale reasoning abilities through structured thought processes.
How It Works
ReasonFlux employs a hierarchical approach, leveraging "thought templates" to guide LLM reasoning. This involves a "navigator" component that selects appropriate templates from a library based on the problem context, and an "inference" model that executes the reasoning steps guided by these templates. This method allows smaller models to achieve performance comparable to or exceeding larger, more general-purpose models on specific reasoning benchmarks.
Quick Start & Requirements
conda create -n ReasonFlux python==3.9, conda activate ReasonFlux, pip install -r requirements.txt).llama-factory for training, lm-evaluation-harness for evaluation, and vllm for inference. Note: Avoid installing flash-attn if using jina-embedding-v3 due to potential conflicts.vllm is also resource-intensive.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
flash-attn and jina-embedding-v3.lm-evaluation-harness framework.1 month ago
1 day
deepseek-ai