Research paper implementation for thought-augmented LLM reasoning
Top 52.4% on sourcepulse
This repository provides the official implementation for "Buffer of Thoughts" (BoT), a novel framework designed to enhance Large Language Model (LLM) reasoning capabilities. It targets researchers and developers working on improving LLM accuracy, efficiency, and robustness in complex reasoning tasks, offering significant performance gains over existing methods.
How It Works
BoT introduces a "meta-buffer" to store distilled "thought-templates" from problem-solving processes. For new problems, relevant templates are retrieved and adaptively instantiated for efficient reasoning. A "buffer-manager" dynamically updates the meta-buffer, ensuring scalability and stability as more tasks are solved. This approach aims to provide superior generalization and robustness while being significantly more cost-effective than multi-query prompting methods.
Quick Start & Requirements
conda create -n BoT python==3.9
), and install dependencies (pip install -r requirements.txt
).python inference.py --api_key 'YOUR_API_KEY'
python run_benchmarks.py --task_name 'benchmark_name' --api_key 'YOUR_API_KEY' --model_id 'MODEL_ID'
Highlighted Details
Maintenance & Community
The project is affiliated with Peking University, UC Berkeley, and Stanford University. Recent updates include the release of ReasonFlux-F1 models and SuperCorrect, a self-correction framework. The implementation of meta-buffer and buffer manager is based on light-RAG.
Licensing & Compatibility
The repository does not explicitly state a license in the README. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The current code primarily supports online LLMs for math problems, with plans to support local models. The README does not detail specific hardware requirements beyond standard Python environments.
1 month ago
1 day