Visuomotor policy learning via action diffusion (research paper)
Top 17.0% on sourcepulse
Diffusion Policy provides a framework for learning visuomotor policies using diffusion models, targeting researchers and engineers in robotics and reinforcement learning. It enables efficient training and evaluation of policies on both simulated and real-world robotic tasks, offering a structured approach to policy learning with state or image-based observations.
How It Works
The core of Diffusion Policy lies in its action-centric diffusion model, which learns to generate a sequence of actions conditioned on a history of observations. It employs a unified interface for tasks and methods, allowing for modularity and extensibility. The framework handles data normalization, policy inference, and training/evaluation orchestration through distinct components like Datasets, Policies, and Workspaces, abstracting away environment-specific details.
Quick Start & Requirements
conda env create -f conda_environment.yaml
or mamba env create -f conda_environment.yaml
.libosmesa6-dev libgl1-mesa-glx libglfw3 patchelf
), RealSense SDK, Spacemouse dependencies (libspnav-dev spacenavd
).Highlighted Details
Maintenance & Community
The project is associated with Columbia University and Toyota Research Institute. Links to experiment logs and further details are available on their website.
Licensing & Compatibility
Limitations & Caveats
The macOS environment setup (conda_environment_macos.yaml
) is noted as having incomplete support for benchmarks. The codebase structure, while flexible, involves code repetition between tasks and methods.
7 months ago
1 day