PyTorch implementation of DDPG for continuous RL
Top 72.0% on sourcepulse
This repository provides a PyTorch implementation of the Deep Deterministic Policy Gradient (DDPG) algorithm for continuous action reinforcement learning problems. It is targeted at researchers and practitioners interested in applying actor-critic methods to control tasks with continuous action spaces. The implementation aims to offer a clear and functional DDPG agent.
How It Works
The core of the implementation is the DDPG algorithm, which uses a deterministic policy for action selection and an Ornstein-Uhlenbeck process for exploration in continuous action spaces. It features separate actor and critic networks, both implemented as 3-layer neural networks. The actor network takes the state as input and outputs an action, while the critic network takes both state and action as input and outputs the state-action value function. Optimization involves minimizing the negative Q-value for the actor and a temporal difference error for the critic, with soft updates applied to target networks to improve stability.
Quick Start & Requirements
pip install -r requirements.txt
Highlighted Details
Maintenance & Community
No specific information on contributors, sponsorships, or community channels is present in the README.
Licensing & Compatibility
The README does not explicitly state a license.
Limitations & Caveats
The README does not detail any specific limitations, known bugs, or deprecation status. The implementation is presented as a functional example rather than a production-ready library.
4 years ago
Inactive