Library for diffusion model training
Top 32.6% on sourcepulse
This library provides scalable and memory-optimized training for diffusion models, targeting researchers and practitioners working with advanced AI video generation. It aims to make complex training algorithms more accessible and efficient.
How It Works
Finetrainers supports distributed training (DDP, FSDP-2, HSDP) and memory-efficient single-GPU training. It offers LoRA and full-rank finetuning, conditional control training, and multiple attention backends (flash, flex, sage, xformers). The library features flexible dataset handling, including combined image/video, chainable local/remote datasets, and multi-resolution bucketing, with memory-efficient precomputation options.
Quick Start & Requirements
pip install -r requirements.txt
and pip install git+https://github.com/huggingface/diffusers
.environment.md
.git fetch --all --tags && git checkout tags/v0.2.0
.Highlighted Details
torch.compile
and multiple attention providers for performance optimization.Maintenance & Community
The project is actively developed with frequent updates. Links to community resources like Discord/Slack are not explicitly provided in the README.
Licensing & Compatibility
The library builds upon and integrates with various open-source libraries. The specific license for finetrainers
itself is not explicitly stated in the README, but its reliance on other libraries implies compatibility with their respective licenses.
Limitations & Caveats
The main development branch is noted as unstable. Some model support (e.g., Wan, CogView4, Flux) has TODO entries for VRAM requirements. HunyuanVideo full finetuning is listed as OOM (Out Of Memory) for the tested configuration.
1 month ago
1 day