Discover and explore top open-source AI tools and projects—updated daily.
naterawVideo generator for Stable Diffusion latent space exploration
Top 10.7% on SourcePulse
This repository provides tools for generating videos by interpolating between text prompts using Stable Diffusion. It's designed for researchers and artists interested in exploring latent space and creating dynamic visual content, offering a novel approach to smooth transitions between generated images.
How It Works
The core functionality is built around the StableDiffusionWalkPipeline, which generates a sequence of images by interpolating through the latent space of a Stable Diffusion model. This allows for smooth visual morphing between different text prompts and seeds. The library also supports synchronizing video generation with audio, using audio cues to control the interpolation speed and create music-driven visuals.
Quick Start & Requirements
pip install stable_diffusion_videostorch_dtype=torch.float16 recommended for NVIDIA GPUs).torch.float32 due to MPS limitations.examples folder.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is presented as a personal project without explicit versioning or formal support. The lack of a specified license may pose restrictions for commercial applications. Apple M1 users need to be aware of the float32 requirement for MPS.
1 year ago
Inactive
Vchitect
Lightricks