RL framework for training LLM agents via end-to-end reinforcement learning
Top 49.4% on sourcepulse
Agent-R1 is an open-source framework for training large language model (LLM) agents using end-to-end reinforcement learning. It targets researchers and developers aiming to build autonomous agents by simplifying the process of defining domain-specific tools and reward functions, eliminating the need for complex workflow engineering. The framework enables agents to learn from complete interaction trajectories, coordinate multiple tools, and process both text and visual inputs.
How It Works
Agent-R1 employs end-to-end reinforcement learning to train agents on entire interaction sequences, allowing for learning from multi-turn tool usage and coordination. It supports custom tool integration via a base class and offers multiple RL algorithms like PPO, GRPO, and REINFORCE++. The framework also incorporates process rewards for individual tool calls, normalized against outcome rewards, and provides multi-modal support through integration with vision-language models.
Quick Start & Requirements
git submodule update --init --recursive
and reinstall verl
locally.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 week ago
Inactive