RL framework for LLM tool use
Top 35.2% on sourcepulse
ReCall is a framework for training Large Language Models (LLMs) to reason and utilize tools via reinforcement learning, without requiring supervised data on tool use. It aims to enable LLMs to act as general-purpose agents by learning to use and combine arbitrary tools.
How It Works
ReCall employs reinforcement learning to train LLMs to interact with tools. It generates synthetic data for training and supports arbitrary tool integration. The framework is built upon a customized version of the verl
reinforcement learning library and utilizes a sandboxed environment for safe execution of tool code.
Quick Start & Requirements
pip3 install -e .
.flash-attn
, and optionally faiss-gpu
for Wikipedia RAG.python sandbox.py --port {port}
.verl
(included in src/verl
) and requires tool URLs, model paths, and W&B API keys. Example training scripts are provided.Highlighted Details
Maintenance & Community
The project is actively being worked on, with the first version released on April 24, 2025. Trained models are available on Hugging Face. The implementation is based on verl
, FlashRAG, BFCL, FastAPI, and SGLang.
Licensing & Compatibility
The repository does not explicitly state a license in the README. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The project is described as a "work in progress." The current sandbox implementation is basic, with plans for a more robust version. The README notes potential security risks with local sandbox hosting.
2 months ago
Inactive