Code for 3D dance generation via Actor-Critic GPT
Top 71.7% on sourcepulse
Bailando is a framework for generating 3D dance sequences from music, targeting researchers and developers in computer vision and animation. It addresses the challenge of creating synchronized, temporally coherent, and spatially constrained dances by leveraging a choreographic memory and an actor-critic GPT model.
How It Works
The core of Bailando is a two-part system: a choreographic memory that quantifies dance units into a codebook, and an actor-critic GPT that composes these units. This approach allows dance generation to operate on quantized units that adhere to choreography standards, ensuring spatial constraints are met. An actor-critic reinforcement learning scheme with a beat-align reward function synchronizes motion tempos with music beats.
Quick Start & Requirements
./prepare_aistpp_data.sh
. Preprocessed features are available.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The reinforcement learning finetuning process for "music in the wild" is not guaranteed to produce satisfying results, and empirical success is observed after <= 30 epochs. The original AIST++ dataset may not cover all music genres or complexities found in real-world dance music.
1 year ago
1 week