AgileRL  by AgileRL

RLOps library for faster reinforcement learning development

created 2 years ago
802 stars

Top 44.9% on sourcepulse

GitHubView on GitHub
Project Summary

AgileRL is a deep reinforcement learning library designed to accelerate RL development through RLOps, focusing on reducing training and hyperparameter optimization (HPO) times. It targets researchers and practitioners seeking efficient RL model development, offering state-of-the-art algorithms and pioneering evolutionary HPO techniques for significant speedups.

How It Works

AgileRL leverages evolutionary algorithms for hyperparameter optimization, automating the discovery of optimal configurations. This approach contrasts with traditional methods requiring numerous manual training runs. The library supports distributed training and includes a range of on-policy, off-policy, offline, multi-agent, and contextual multi-armed bandit algorithms, all designed to be "evolvable."

Quick Start & Requirements

Highlighted Details

  • Claims 10x faster hyperparameter optimization compared to traditional RL frameworks combined with Optuna.
  • Supports multi-agent RL using Petting Zoo-style API, with benchmarks against other libraries.
  • Implements a wide array of algorithms including PPO, DQN, Rainbow DQN, DDPG, TD3, CQL, ILQL, MADDPG, MATD3, NeuralUCB, and NeuralTS.
  • Features include custom module/network creation and LLM finetuning capabilities.

Maintenance & Community

Licensing & Compatibility

  • License: Apache 2.0.
  • Compatibility: Permissive license suitable for commercial use and integration with closed-source projects.

Limitations & Caveats

The library is under active development, with "more coming soon" for evolvable algorithms. While benchmarks claim significant speedups, real-world performance may vary based on specific environments and configurations.

Health Check
Last commit

4 days ago

Responsiveness

1 day

Pull Requests (30d)
24
Issues (30d)
7
Star History
58 stars in the last 90 days

Explore Similar Projects

Starred by Lewis Tunstall Lewis Tunstall(Researcher at Hugging Face), Robert Nishihara Robert Nishihara(Cofounder of Anyscale; Author of Ray), and
4 more.

verl by volcengine

2.4%
12k
RL training library for LLMs
created 9 months ago
updated 16 hours ago
Feedback? Help us improve.