PyTorch library for deep reinforcement learning research
Top 20.4% on sourcepulse
This repository provides modular, optimized implementations of common deep reinforcement learning algorithms in PyTorch, targeting researchers and practitioners for small- to medium-scale experiments. It offers a unified infrastructure for policy gradient, deep Q-learning, and Q-function policy gradient methods, enabling high-throughput research with flexible parallelization and multi-GPU support.
How It Works
The library is built around a modular design, separating concerns into components like Runner
, Sampler
, Collector
, Agent
, Model
, and Algorithm
. It utilizes a custom namedarraytuple
data structure for efficient organization and manipulation of NumPy arrays and PyTorch tensors, simplifying handling of multi-modal observations and actions. This approach allows for seamless integration of different components and facilitates easy modification and extension of algorithms.
Quick Start & Requirements
linux_[cpu|cuda9|cuda10].yml
). Then, either add the rlpyt
directory to PYTHONPATH
or install it as an editable package (pip install -e .
).Highlighted Details
DistributedDataParallel
.namedarraytuple
for efficient data handling, supporting multi-modal observations/actions.Maintenance & Community
The project acknowledges support from Pieter Abbeel, the Fannie & John Hertz Foundation, NVIDIA, Max Jaderberg, OpenAI, and the BAIR community. Contributions are welcomed.
Licensing & Compatibility
The repository does not explicitly state a license in the provided README. Compatibility with commercial or closed-source linking is not specified.
Limitations & Caveats
The README indicates that the code is stable but may still develop, with potential for future changes. Some algorithms are listed as "Coming soon." The project does not include its own visualization tools, recommending https://github.com/vitchyr/viskit
instead.
4 years ago
Inactive