Self-RAG implementation for learning retrieval, generation, and critique via self-reflection
Top 21.4% on sourcepulse
Self-RAG is a framework for training Large Language Models (LLMs) to retrieve, generate, and critique text, enhancing factual accuracy and quality. It targets researchers and developers aiming to improve LLM factuality without sacrificing versatility, offering on-demand retrieval and self-critique capabilities.
How It Works
Self-RAG integrates retrieval and critique as integral parts of the generation process. The model learns to predict "reflection tokens" to assess retrieved passages and its own generations across multiple fine-grained aspects. A segment-wise beam search then selects outputs that maximize user-defined preferences, allowing for dynamic retrieval or skipping based on query needs.
Quick Start & Requirements
pip install -r requirements.txt
vllm
library (ensure latest version for skip_special_tokens
parameter).selfrag/selfrag_llama2_7b
).Highlighted Details
Maintenance & Community
SciPhi-Self-RAG-Mistral-7B-32k
.Licensing & Compatibility
Limitations & Caveats
1 year ago
1 week