Open-source framework for decentralized RL training swarms
Top 33.1% on sourcepulse
RL Swarm is an open-source, peer-to-peer framework for distributed reinforcement learning training. It enables users to collaboratively train large language models by connecting their hardware to a decentralized network, leveraging collective intelligence for faster and more efficient model development. The system is designed for both consumer laptops and cloud GPUs, offering permissionless participation.
How It Works
RL Swarm utilizes a peer-to-peer architecture where individual nodes contribute computational resources to train models collaboratively. It supports various Qwen 2.5 models and datasets, allowing users to select configurations based on their hardware capabilities. The system includes an optional on-chain identity management layer via Alchemy for tracking progress and participation.
Quick Start & Requirements
./run_rl_swarm.sh
after setting up a Python virtual environment (python3 -m venv .venv
, source .venv/bin/activate
).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
This software is experimental and provided as-is. Performance on consumer hardware may be slow, and some configurations or platforms (e.g., certain VPSs, Windows without WSL) may require significant debugging. On-chain identity management has specific requirements regarding swarm.pem
and email registration to function correctly.
1 week ago
1 day