Stable diffusion training/fine-tuning codebase
Top 90.9% on sourcepulse
This repository provides a Stable Diffusion implementation back-ported to OpenAI's guided diffusion codebase, enabling easier development and training of diffusion models. It targets researchers and developers interested in advanced image generation techniques like inpainting, outpainting, classifier guidance, and super-resolution.
How It Works
The project leverages the established architecture of OpenAI's guided diffusion, integrating Stable Diffusion's components. This approach allows for modularity and extensibility, facilitating experimentation with various sampling methods, guidance scales, and model configurations. The use of MPI for distributed training enables efficient scaling across multiple GPUs.
Quick Start & Requirements
pip install latent-diffusion
), then install this project (pip install -e .
). Install MPI and mpi4py
(sudo apt install libopenmpi-dev
, pip install mpi4py
).diffusion.pt
and kl.pt
from Hugging Face or split a Stable Diffusion checkpoint using split.py
.mpi4py
, PyQt5
(for GUI).Highlighted Details
PyQt5
.Maintenance & Community
The repository appears to be a personal project by Jack000. No specific community channels or active development signals are present in the README.
Licensing & Compatibility
The README does not explicitly state a license. The project depends on Latent Diffusion, which is typically under permissive licenses (e.g., MIT). Compatibility with commercial or closed-source projects would require license verification.
Limitations & Caveats
Training scripts specify high VRAM requirements, potentially limiting accessibility for users without high-end hardware. The project's reliance on the older OpenAI guided diffusion codebase might mean it misses recent advancements or optimizations found in more current Stable Diffusion implementations.
2 years ago
Inactive