Discover and explore top open-source AI tools and projects—updated daily.
Reinforcement learning infrastructure for agentic AI
Top 73.9% on SourcePulse
RLinf is an open-source infrastructure designed for post-training foundation models (LLMs, VLMs, VLAs) using reinforcement learning. It provides a flexible and scalable backbone for developing agentic AI, enabling open-ended learning and continuous generalization. The system is particularly beneficial for researchers and developers working on advanced AI training paradigms.
How It Works
RLinf introduces a novel "Macro-to-Micro Flow" (M2Flow) paradigm, which separates the logical workflow construction from physical communication and scheduling. This allows for programmable, high-level logical flows to be executed efficiently through micro-level operations. It supports flexible execution modes (Collocated, Disaggregated, Hybrid) and an automatic scheduling strategy that selects the optimal mode based on the training workload, eliminating the need for manual resource allocation.
Quick Start & Requirements
Highlighted Details
Maintenance & Community
RLinf is a new project, with its formal v0.1 release and accompanying paper expected soon. It acknowledges inspiration from projects like VeRL, AReaL, Megatron-LM, SGLang, and PyTorch FSDP. Contact information for inquiries and potential collaborators is provided.
Licensing & Compatibility
The README does not explicitly state the license type or compatibility for commercial use.
Limitations & Caveats
The project is in its early stages, with a formal v0.1 release and paper forthcoming. The roadmap indicates planned support for heterogeneous GPUs, asynchronous pipeline execution, Mixture of Experts (MoE), vLLM inference backend, and various VLM/VLA training extensions, suggesting these features are not yet available.
21 hours ago
Inactive