Video stylization tool using StableDiffusion and ControlNet
Top 44.1% on sourcepulse
This project provides an extension for the Automatic1111 Stable Diffusion web UI to automate video stylization and generation. It targets users looking to create stylized videos from existing footage (vid2vid) or generate entirely new videos from text prompts, offering control over resolution and length. The key benefit is enhanced stability and quality in video generation through optical flow estimation.
How It Works
The extension leverages RAFT for optical flow estimation in vid2vid mode to maintain animation stability and generate occlusion masks for frame-to-frame consistency. For text-to-video generation, it utilizes a "FloweR" method (in progress) to predict optical flow. ControlNet integration is crucial, especially in vid2vid, to prevent choppy results. The text-to-video mode also supports using video as guidance for ControlNet, enabling stronger stylization.
Quick Start & Requirements
https://github.com/volotat/SD-CN-Animation.git
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
1 day