Discover and explore top open-source AI tools and projects—updated daily.
Neural network structure for adding conditional control to diffusion models
Top 1.0% on SourcePulse
ControlNet provides a method to add conditional control to diffusion models, enabling fine-grained manipulation of image generation based on various inputs like edge maps, depth maps, or human poses. It's designed for researchers and artists looking to precisely guide text-to-image synthesis without compromising pre-trained diffusion models.
How It Works
ControlNet achieves control by duplicating diffusion model weights into a "locked" (original) and a "trainable" copy. The trainable copy learns the conditioning input via a "zero convolution" layer, which initially outputs zeros, preventing distortion. This architecture allows training on small datasets while preserving the integrity of the original, powerful diffusion model backbone.
Quick Start & Requirements
conda env create -f environment.yaml
and conda activate control
.python gradio_*.py
scripts (e.g., gradio_canny2image.py
).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
Inactive