PyTorch image-to-image diffusion model implementation
Top 25.4% on sourcepulse
This repository provides an unofficial PyTorch implementation of Palette: Image-to-Image Diffusion Models, targeting researchers and practitioners in generative AI. It offers a framework for various image-to-image tasks like inpainting, uncropping, and colorization, leveraging a U-Net architecture and attention mechanisms for enhanced sample quality.
How It Works
The implementation adapts the U-Net architecture from Guided-Diffusion, incorporating attention mechanisms in low-resolution features (16x16) similar to DDPM. It encodes the conditioning signal $\gamma$ directly, embedding it via affine transformation, and fixes the variance $\Sigma_\theta(x_t, t)$ to a constant during inference, as described in the Palette paper. This approach aims for robust performance and high-quality image generation.
Quick Start & Requirements
pip install -r requirements.txt
.Highlighted Details
Maintenance & Community
The project is an unofficial implementation and does not list specific maintainers or community channels. It acknowledges inspiration from OpenAI's guided-diffusion
and Diffusion-Based-Model-for-Colorization
.
Licensing & Compatibility
The repository's license is not explicitly stated in the README. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The project is an unofficial implementation and notes that follow-up experiments are uncertain due to time and GPU resource constraints. Some tasks like uncropping and colorization are marked as not yet implemented. The DDPM model requires significant computational resources.
2 years ago
Inactive