Starter agent for solving Universe environments using A3C
Top 35.3% on sourcepulse
This repository provides a starter agent for solving various "universe" environments, primarily targeting researchers and developers in reinforcement learning. It implements a basic Actor-Critic algorithm (A3C) adapted for real-time, potentially high-latency environments, offering a foundational example for building more complex agents.
How It Works
The agent utilizes an asynchronous advantage actor-critic (A3C) algorithm, a popular reinforcement learning technique known for its efficiency in parallel training. It spawns multiple worker processes that interact with the environment and update a central parameter server. This asynchronous approach allows for continuous learning without waiting for individual workers to complete, making it suitable for real-time environments where latency is a significant factor.
Quick Start & Requirements
conda create --name universe-starter-agent python=3.5
, source activate universe-starter-agent
, then install dependencies via pip
and conda
as detailed in the README.six
, TensorFlow 0.12, tmux
, htop
, cmake
, libjpeg-turbo
(or libjpeg-dev
on Linux), gym[atari]
, universe
, opencv-python
, numpy
, scipy
.Highlighted Details
env.render()
.Maintenance & Community
This repository has been deprecated in favor of the Retro library.
Licensing & Compatibility
The repository's license is not explicitly stated in the README.
Limitations & Caveats
The project is deprecated. Performance is highly sensitive to network latency, especially for real-time environments. The provided implementation is tuned for VNC Pong and may require significant adjustments for other tasks.
7 years ago
Inactive