DDPG implementation for continuous control in OpenAI Gym
Top 94.7% on sourcepulse
This repository provides a TensorFlow implementation of the Deep Deterministic Policy Gradient (DDPG) algorithm for continuous control tasks, targeting researchers and practitioners in reinforcement learning. It aims to offer a clear and functional DDPG agent for OpenAI Gym environments.
How It Works
The implementation follows the DDPG algorithm as described in Lillicrap et al. (arXiv:1509.02971). It utilizes an actor-critic architecture with separate networks for the policy and value functions. Key features include batch normalization for improved learning speed and a "grad-inverter" component, referencing arXiv:1511.04143.
Quick Start & Requirements
git clone https://github.com/stevenpjg/ddpg-aigym.git
and cd ddpg-aigym
.python main.py
.Highlighted Details
Maintenance & Community
No information on maintainers, community channels, or roadmap is provided in the README.
Licensing & Compatibility
The README does not specify a license. Compatibility for commercial use or closed-source linking is undetermined.
Limitations & Caveats
The implementation is based on an older TensorFlow version (0.11.0rc0), which may present compatibility issues with modern TensorFlow or Python versions. The lack of explicit licensing information is a significant caveat for adoption.
7 years ago
Inactive