ControlNet-v1-1-nightly  by lllyasviel

ControlNet 1.1 is a research project for neural network control of diffusion models

Created 2 years ago
5,089 stars

Top 9.8% on SourcePulse

GitHubView on GitHub
Project Summary

This repository provides the nightly release of ControlNet 1.1, a powerful extension for Stable Diffusion that enables fine-grained control over image generation using various conditioning inputs. It's primarily aimed at researchers and users who want to leverage advanced image-to-image translation and manipulation capabilities within the Stable Diffusion ecosystem.

How It Works

ControlNet 1.1 maintains the same architecture as its predecessor, focusing on improved robustness and quality across its suite of 14 models (11 production-ready, 3 experimental). These models condition Stable Diffusion generation on inputs like depth maps, Canny edges, segmentation maps, OpenPose skeletons, and more. The key advantage lies in its ability to inject spatial conditioning into the diffusion process, allowing for precise control over composition, structure, and style.

Quick Start & Requirements

  • Installation: Clone the repository and set up the environment using conda env create -f environment.yaml followed by conda activate control-v11.
  • Dependencies: Requires Python, PyTorch, and specific annotator models (which can be downloaded automatically or manually). A Stable Diffusion 1.5 checkpoint (v1-5-pruned.ckpt) is also needed.
  • GPU: An 8GB GPU requires setting save_memory = True in config.py.
  • Usage: Examples for specific models are provided via python gradio_*.py scripts. For Automatic1111 Stable Diffusion WebUI integration, users should refer to the sd-webui-controlnet repository.

Highlighted Details

  • Offers 14 models including Canny, Depth, Normal, OpenPose, Scribble, SoftEdge, Segmentation, MLSD, Lineart, Shuffle, Inpaint, and IP2P.
  • Introduces "SoftEdge_safe" to mitigate artifacts from HED/PIDI preprocessors.
  • Normal model 1.1 is significantly improved for real-world normal map compatibility.
  • OpenPose 1.1 features improved accuracy, especially for hands, and supports body, hand, and face inputs.
  • Shuffle model 1.1 is positioned as the sole maintained method for image stylization.

Maintenance & Community

This repository is actively being updated. For integration with Automatic1111's WebUI, users should follow the sd-webui-controlnet repository.

Licensing & Compatibility

The repository does not explicitly state a license in the provided README. Users should verify licensing for commercial use.

Limitations & Caveats

This repository is for research use and academic experiments; it is not an A1111 extension and should not be directly installed into A1111. Official support for Multi-ControlNet is A1111-only. The "Instruct Pix2Pix" model is marked as experimental and may require cherry-picking. The "Tile" model's official support for tiled upscaling is A1111-only.

Health Check
Last Commit

1 year ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
0
Star History
23 stars in the last 30 days

Explore Similar Projects

Starred by Chip Huyen Chip Huyen(Author of "AI Engineering", "Designing Machine Learning Systems"), Luca Antiga Luca Antiga(CTO of Lightning AI), and
2 more.

mmagic by open-mmlab

0.1%
7k
AIGC toolbox for image/video editing and generation
Created 6 years ago
Updated 1 year ago
Feedback? Help us improve.