Discover and explore top open-source AI tools and projects—updated daily.
Jax diffusion models for training and inference
Top 97.5% on SourcePulse
MaxDiffusion is a collection of pure Python/Jax implementations for latent diffusion models, optimized for XLA devices like Cloud TPUs and GPUs. It serves as a research and production launching point for ambitious diffusion projects, enabling users to train, tune, and serve solutions with models like Stable Diffusion 2.x, XL, Flux, and LTX-Video.
How It Works
MaxDiffusion leverages Jax and XLA for high-performance, distributed computing across TPU pods. Its architecture is designed for scalability and efficiency, allowing for complex operations like fused attention (via Transformer Engine on GPUs) and multi-host training. The project supports various diffusion models and offers features like LoRA loading and ControlNet inference, providing a flexible foundation for advanced diffusion tasks.
Quick Start & Requirements
pip install -r requirements.txt
, pip install .
(after cloning). For GPU with fused attention: pip install -U "jax[cuda12]"
, pip install "transformer_engine[jax]"
.gcloud
.Highlighted Details
Maintenance & Community
The project is hosted on GitHub at AI-Hypercomputer/maxdiffusion. Community interaction details (Discord/Slack, etc.) are not specified in the README.
Licensing & Compatibility
The README does not explicitly state a license. Compatibility with Hugging Face Jax models is noted.
Limitations & Caveats
Flux finetuning and some LoRA formats have limited testing. Specific hardware configurations (e.g., TPU v5p for Flux finetuning) are recommended or tested. The README does not detail potential bus factors or provide a roadmap.
1 day ago
Inactive