Agentic RAG framework for large reasoning models
Top 38.1% on sourcepulse
Search-o1 is a framework designed to enhance Large Reasoning Models (LRMs) by integrating an agentic Retrieval-Augmented Generation (RAG) mechanism. It addresses the knowledge insufficiency and uncertainty issues that LRMs face during prolonged reasoning processes, making it suitable for researchers and developers working on complex, knowledge-intensive tasks.
How It Works
Search-o1 augments LRMs with an agentic search workflow and a Reason-in-Documents module. The agentic search allows the model to dynamically retrieve external knowledge when encountering uncertain information. The Reason-in-Documents module then refines and integrates this retrieved information into the reasoning chain, reducing noise and maintaining coherence. This is achieved through a batch generation mechanism with interleaved search, where queries are detected and documents are retrieved, refined, and integrated in batches to improve reliability and accuracy.
Quick Start & Requirements
conda create -n search_o1 python=3.9
) and install requirements (pip install -r requirements.txt
).data/data_pre_process.ipynb
. Supported datasets include GPQA, MATH500, AMC2023, AIME2024, LiveCodeBench, NQ, TriviaQA, HotpotQA, 2WikiMultihopQA, MuSiQue, and Bamboogle.scripts/run_search_o1.py
.Highlighted Details
Maintenance & Community
The project is associated with authors from Renmin University of China. Contact is available via email at xiaoxi_li@ruc.edu.cn. A "To Do List" indicates ongoing development for adding more backbone models, improving tool utilization, and incorporating additional tools.
Licensing & Compatibility
Limitations & Caveats
The evaluation strategy includes a backoff to direct generation if retrieval-based methods fail to provide a final answer, suggesting potential issues with the effective utilization of retrieved text by the model in its current state.
2 months ago
1 day