RL library for large language models
Top 26.8% on sourcepulse
ROLL is an open-source library designed to scale Reinforcement Learning (RL) for Large Language Models (LLMs) using distributed GPU resources. It targets AI labs, hyperscalers, and product developers aiming to enhance LLM capabilities in areas like human preference alignment, complex reasoning, and agentic interactions, offering significant speedups and cost reductions.
How It Works
ROLL employs a multi-role distributed architecture, leveraging Ray for flexible resource allocation and heterogeneous task scheduling. It integrates with high-performance backends such as Megatron-Core, SGLang, and vLLM to accelerate training and inference. The library emphasizes efficient data handling, including sample filtering based on difficulty and length, and provides advanced techniques for stabilizing training, such as value/advantage clipping and reward normalization.
Quick Start & Requirements
pip install -e .
(from source)Highlighted Details
Maintenance & Community
Developed by Alibaba TAOBAO & TMALL Group and Alibaba Group. The project actively posts updates and has a tech report available. Community contributions are welcomed.
Licensing & Compatibility
Licensed under the Apache License (Version 2.0). The project utilizes third-party components under other open-source licenses, as detailed in the NOTICE file. Compatible with commercial use.
Limitations & Caveats
The project is actively under development with upcoming features like Qwen2.5 VL RL pipeline and FSDP2 integration. While it supports single-GPU setups, its primary design focus is on large-scale GPU clusters.
20 hours ago
Inactive