RL via sequence modeling research paper
Top 18.3% on sourcepulse
This repository provides the official codebase for Decision Transformer, a method that frames Reinforcement Learning (RL) as a sequence modeling problem. It is intended for researchers and practitioners in RL and deep learning who are interested in applying transformer architectures to sequential decision-making tasks. The primary benefit is enabling RL agents to learn from historical trajectories using standard sequence modeling techniques.
How It Works
Decision Transformer leverages the transformer architecture, commonly used in Natural Language Processing, to model sequences of states, actions, and rewards. Instead of traditional RL algorithms that rely on value functions or policy gradients, it treats RL as a conditional sequence generation problem. The model predicts future actions based on a sequence of past states, actions, and desired future returns, effectively learning a policy that aims to achieve a specified level of performance.
Quick Start & Requirements
atari
or gym
subdirectories. Add respective directories to PYTHONPATH
.Highlighted Details
atari
and gym
for distinct experiment types.Maintenance & Community
The project is associated with authors from leading research institutions. No specific community channels or active maintenance signals are provided in the README.
Licensing & Compatibility
Limitations & Caveats
The README states this is not an official Google or Facebook product. The codebase is primarily for reproducing paper experiments, and may require significant effort to adapt for novel applications or different RL environments.
1 year ago
Inactive