Research paper for distributed parallel inference of high-resolution diffusion models
Top 49.8% on sourcepulse
DistriFusion addresses the challenge of accelerating high-resolution diffusion model inference across multiple GPUs without compromising image quality. It is designed for researchers and practitioners working with large-scale generative models who need to reduce latency and improve throughput for tasks like text-to-image generation.
How It Works
DistriFusion employs a training-free distributed inference strategy by partitioning the image generation process across multiple GPUs. It utilizes a novel synchronous communication approach for patch interaction in the initial step, followed by asynchronous communication to reuse activations from previous steps. This technique effectively hides communication overhead within the computation pipeline, enabling significant speedups.
Quick Start & Requirements
pip install distrifuser
or pip install git+https://github.com/mit-han-lab/distrifuser.git
torchrun --nproc_per_node=$N_GPUS scripts/sdxl_example.py
.Highlighted Details
diffusers
library.Maintenance & Community
Licensing & Compatibility
diffusers
and lmxyy/sige
.Limitations & Caveats
8 months ago
1 day