chainerrl  by chainer

Deep RL library for algorithm experimentation

created 8 years ago
1,197 stars

Top 33.4% on sourcepulse

GitHubView on GitHub
Project Summary

ChainerRL is a Python library for deep reinforcement learning, offering a comprehensive suite of state-of-the-art algorithms and techniques. It targets researchers and practitioners in RL, providing a flexible framework built on Chainer for developing and experimenting with agents.

How It Works

ChainerRL implements a wide array of RL algorithms, including DQN variants, DDPG, A3C, PPO, and SAC, supporting both discrete and continuous action spaces, recurrent models, and batch/asynchronous training where applicable. It leverages Chainer's flexibility for defining neural network architectures and training loops, enabling efficient implementation and customization of RL agents.

Quick Start & Requirements

Highlighted Details

  • Implements advanced techniques like NoisyNet, Prioritized Experience Replay, Dueling Networks, and Normalized Advantage Function.
  • Supports visualization tools for inspecting and debugging agent behavior.
  • Compatible with any environment adhering to the OpenAI Gym interface.
  • Offers implementations for both synchronous (A2C) and asynchronous (A3C) training variants.

Maintenance & Community

The project is associated with the Chainer deep learning framework. Further community engagement details are not explicitly provided in the README.

Licensing & Compatibility

  • License: MIT License.
  • Compatibility: Permissive MIT license allows for commercial use and integration with closed-source projects.

Limitations & Caveats

The library is built on Chainer, which has been succeeded by CuPy and PyTorch. While ChainerRL itself is functional, the underlying framework's development status may impact long-term support and integration with newer deep learning ecosystems.

Health Check
Last commit

4 years ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
11 stars in the last 90 days

Explore Similar Projects

Feedback? Help us improve.