Neural network structure for adding conditional control to diffusion models
Top 1.1% on sourcepulse
ControlNet provides a method to add conditional control to diffusion models, enabling fine-grained manipulation of image generation based on various inputs like edge maps, depth maps, or human poses. It's designed for researchers and artists looking to precisely guide text-to-image synthesis without compromising pre-trained diffusion models.
How It Works
ControlNet achieves control by duplicating diffusion model weights into a "locked" (original) and a "trainable" copy. The trainable copy learns the conditioning input via a "zero convolution" layer, which initially outputs zeros, preventing distortion. This architecture allows training on small datasets while preserving the integrity of the original, powerful diffusion model backbone.
Quick Start & Requirements
conda env create -f environment.yaml
and conda activate control
.python gradio_*.py
scripts (e.g., gradio_canny2image.py
).Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
1 year ago
Inactive