Research paper for video reconstruction from brain activity
Top 75.9% on sourcepulse
MinD-Video is a framework for reconstructing high-quality videos from fMRI brain activity data, targeting researchers and engineers in neuroscience and AI. It enables the visualization of visual experiences directly from brain recordings, advancing the understanding of cognitive processes.
How It Works
MinD-Video employs a multi-stage approach: masked brain modeling to learn spatiotemporal patterns from fMRI data, multimodal contrastive learning with spatiotemporal attention for robust feature extraction, and co-training with a Stable Diffusion model enhanced by temporal inflation. This combination allows for high-quality, arbitrary frame rate video generation guided by adversarial principles.
Quick Start & Requirements
conda env create -f env.yaml
and conda activate mind-video
.python scripts/eval_all.py --config configs/eval_all_sub1.yaml
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
Inactive