Video generator for Stable Diffusion latent space exploration
Top 10.9% on sourcepulse
This repository provides tools for generating videos by interpolating between text prompts using Stable Diffusion. It's designed for researchers and artists interested in exploring latent space and creating dynamic visual content, offering a novel approach to smooth transitions between generated images.
How It Works
The core functionality is built around the StableDiffusionWalkPipeline
, which generates a sequence of images by interpolating through the latent space of a Stable Diffusion model. This allows for smooth visual morphing between different text prompts and seeds. The library also supports synchronizing video generation with audio, using audio cues to control the interpolation speed and create music-driven visuals.
Quick Start & Requirements
pip install stable_diffusion_videos
torch_dtype=torch.float16
recommended for NVIDIA GPUs).torch.float32
due to MPS limitations.examples
folder.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is presented as a personal project without explicit versioning or formal support. The lack of a specified license may pose restrictions for commercial applications. Apple M1 users need to be aware of the float32
requirement for MPS.
10 months ago
1 day