ICCV 2023 paper implementing training-free conditional diffusion
Top 89.3% on sourcepulse
FreeDoM is a training-free method for controlling unconditional diffusion models using various conditions like text, sketches, and face IDs. It enables conditional generation across different domains, including human faces and ImageNet, by leveraging pre-trained networks to guide the diffusion process, offering a flexible approach to controlled image synthesis.
How It Works
FreeDoM constructs a time-independent energy function using off-the-shelf pre-trained networks. This function quantifies the discrepancy between intermediate generated images and desired conditions. By computing the gradient of this energy function, FreeDoM guides the diffusion sampling process, allowing for condition-specific generation without requiring model fine-tuning.
Quick Start & Requirements
pip install -r requirements.txt
.Highlighted Details
Maintenance & Community
The project is the official implementation for an ICCV 2023 paper. The README indicates ongoing development with completed tasks including code release for human face models, ControlNet integration, and Stable Diffusion style guidance. No specific community links (Discord, Slack) are provided.
Licensing & Compatibility
The repository's license is not explicitly stated in the README. However, it heavily relies on and acknowledges other open-source projects like SDEdit, guided-diffusion, Stable Diffusion, and ControlNet, which have their own licenses. Compatibility for commercial use would require verifying the licenses of all dependencies.
Limitations & Caveats
The README does not detail specific limitations or known bugs. The sampling times are reported on a single GPU (RTX 3090), and performance on different hardware may vary. The project is presented as an official implementation of a research paper, suggesting it may primarily focus on research use cases.
1 year ago
Inactive