ControlNet-v1-1-nightly  by lllyasviel

ControlNet 1.1 is a research project for neural network control of diffusion models

created 2 years ago
5,056 stars

Top 10.1% on sourcepulse

GitHubView on GitHub
Project Summary

This repository provides the nightly release of ControlNet 1.1, a powerful extension for Stable Diffusion that enables fine-grained control over image generation using various conditioning inputs. It's primarily aimed at researchers and users who want to leverage advanced image-to-image translation and manipulation capabilities within the Stable Diffusion ecosystem.

How It Works

ControlNet 1.1 maintains the same architecture as its predecessor, focusing on improved robustness and quality across its suite of 14 models (11 production-ready, 3 experimental). These models condition Stable Diffusion generation on inputs like depth maps, Canny edges, segmentation maps, OpenPose skeletons, and more. The key advantage lies in its ability to inject spatial conditioning into the diffusion process, allowing for precise control over composition, structure, and style.

Quick Start & Requirements

  • Installation: Clone the repository and set up the environment using conda env create -f environment.yaml followed by conda activate control-v11.
  • Dependencies: Requires Python, PyTorch, and specific annotator models (which can be downloaded automatically or manually). A Stable Diffusion 1.5 checkpoint (v1-5-pruned.ckpt) is also needed.
  • GPU: An 8GB GPU requires setting save_memory = True in config.py.
  • Usage: Examples for specific models are provided via python gradio_*.py scripts. For Automatic1111 Stable Diffusion WebUI integration, users should refer to the sd-webui-controlnet repository.

Highlighted Details

  • Offers 14 models including Canny, Depth, Normal, OpenPose, Scribble, SoftEdge, Segmentation, MLSD, Lineart, Shuffle, Inpaint, and IP2P.
  • Introduces "SoftEdge_safe" to mitigate artifacts from HED/PIDI preprocessors.
  • Normal model 1.1 is significantly improved for real-world normal map compatibility.
  • OpenPose 1.1 features improved accuracy, especially for hands, and supports body, hand, and face inputs.
  • Shuffle model 1.1 is positioned as the sole maintained method for image stylization.

Maintenance & Community

This repository is actively being updated. For integration with Automatic1111's WebUI, users should follow the sd-webui-controlnet repository.

Licensing & Compatibility

The repository does not explicitly state a license in the provided README. Users should verify licensing for commercial use.

Limitations & Caveats

This repository is for research use and academic experiments; it is not an A1111 extension and should not be directly installed into A1111. Official support for Multi-ControlNet is A1111-only. The "Instruct Pix2Pix" model is marked as experimental and may require cherry-picking. The "Tile" model's official support for tiled upscaling is A1111-only.

Health Check
Last commit

11 months ago

Responsiveness

1+ week

Pull Requests (30d)
0
Issues (30d)
0
Star History
88 stars in the last 90 days

Explore Similar Projects

Starred by Tobi Lutke Tobi Lutke(Cofounder of Shopify) and Patrick von Platen Patrick von Platen(Core Contributor to Hugging Face Transformers and Diffusers).

taesd by madebyollin

0.5%
758
Tiny AutoEncoder for Stable Diffusion latents
created 2 years ago
updated 3 months ago
Starred by Lilian Weng Lilian Weng(Cofounder of Thinking Machines Lab), Patrick Kidger Patrick Kidger(Core Contributor to JAX ecosystem), and
4 more.

glow by openai

0.1%
3k
Generative flow research paper code
created 7 years ago
updated 1 year ago
Feedback? Help us improve.