decision-transformer  by kzl

RL via sequence modeling research paper

created 4 years ago
2,618 stars

Top 18.3% on sourcepulse

GitHubView on GitHub
Project Summary

This repository provides the official codebase for Decision Transformer, a method that frames Reinforcement Learning (RL) as a sequence modeling problem. It is intended for researchers and practitioners in RL and deep learning who are interested in applying transformer architectures to sequential decision-making tasks. The primary benefit is enabling RL agents to learn from historical trajectories using standard sequence modeling techniques.

How It Works

Decision Transformer leverages the transformer architecture, commonly used in Natural Language Processing, to model sequences of states, actions, and rewards. Instead of traditional RL algorithms that rely on value functions or policy gradients, it treats RL as a conditional sequence generation problem. The model predicts future actions based on a sequence of past states, actions, and desired future returns, effectively learning a policy that aims to achieve a specified level of performance.

Quick Start & Requirements

  • Install: Clone the repository and follow instructions within the atari or gym subdirectories. Add respective directories to PYTHONPATH.
  • Prerequisites: Python, PyTorch, OpenAI Gym, Atari environments. Specific dependencies are detailed in the sub-directory READMEs.
  • Resources: Requires significant computational resources for training, especially for Atari experiments.
  • Links: arXiv Paper

Highlighted Details

  • Reproduces experiments from the Decision Transformer paper.
  • Codebase split into atari and gym for distinct experiment types.
  • Applies transformer architecture to RL problems.

Maintenance & Community

The project is associated with authors from leading research institutions. No specific community channels or active maintenance signals are provided in the README.

Licensing & Compatibility

  • License: MIT License.
  • Compatibility: Permissive license allows for commercial use and integration into closed-source projects.

Limitations & Caveats

The README states this is not an official Google or Facebook product. The codebase is primarily for reproducing paper experiments, and may require significant effort to adapt for novel applications or different RL environments.

Health Check
Last commit

1 year ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
71 stars in the last 90 days

Explore Similar Projects

Starred by Jiayi Pan Jiayi Pan(Author of SWE-Gym; AI Researcher at UC Berkeley), Tom Brown Tom Brown(Cofounder of Anthropic), and
1 more.

spinningup by openai

0.2%
11k
Educational resource for learning deep reinforcement learning
created 6 years ago
updated 1 year ago
Feedback? Help us improve.