Universal control framework for diffusion transformer models
Top 25.5% on sourcepulse
OminiControl provides a minimal and universal framework for controlling Diffusion Transformer models, specifically FLUX.1. It enables subject-driven and spatial control (e.g., edge-guided, in-painting) with minimal parameter overhead, making it suitable for researchers and developers looking to enhance generative AI capabilities.
How It Works
OminiControl injects control signals into Diffusion Transformers with a minimal design, adding only ~0.1% to the base model's parameters. This approach preserves the original model architecture while enabling diverse control mechanisms, including subject replication and spatial conditioning like edge-to-image or in-painting.
Quick Start & Requirements
pip install -r requirements.txt
Highlighted Details
Maintenance & Community
The project has released OminiControl2 with efficient conditioning methods and supports custom style LoRAs. Training code and higher-resolution models have been released. Links to demos and inference examples are provided.
Licensing & Compatibility
The README does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
Subject-driven generation is primarily optimized for objects, not human subjects, due to training data limitations. The released models currently support only 512x512 resolution for subject-driven generation, though 1024x1024 models are mentioned as released. The subject-driven model may not perform well with FLUX.1-dev.
1 month ago
Inactive