Discover and explore top open-source AI tools and projects—updated daily.
Accelerate diffusion transformer inference with unified caching
Top 85.4% on SourcePulse
cache-dit
is a Python toolbox accelerating Diffusion Transformer (DiT) models within 🤗Diffusers. It offers training-free cache acceleration via techniques like DBCache and TaylorSeer, significantly speeding up inference. Targeting researchers and engineers, it provides a unified API for easy integration across numerous DiT architectures.
How It Works
The library reduces redundant computations using caching mechanisms. Its unified API (cache_dit.enable_cache
) simplifies integration. Key techniques include DBCache, which balances performance and precision through configurable blocks (Fn, Bn), and Hybrid TaylorSeer for improved accuracy with larger cache steps using Taylor series expansion. It also supports CFG caching and torch.compile
compatibility.
Quick Start & Requirements
Install via pip: pip install -U cache-dit
. Requires Python, 🤗Diffusers, and PyTorch. GPU acceleration is recommended. Repository examples and documentation detail integration for specific models.
Highlighted Details
torch.compile
.Maintenance & Community
Primarily associated with "vipshop.com". Community contribution is encouraged via GitHub stars and CONTRIBUTE.md
. No specific community channels or roadmap details are provided.
Licensing & Compatibility
The license type is not specified in the provided README. Compatible with 🤗Diffusers and torch.compile
.
Limitations & Caveats
Unified cache APIs are experimental. torch.compile
with dynamic shapes may require torch._dynamo
recompile limit adjustments. Project authorship appears concentrated, potentially indicating a low bus factor.
20 hours ago
Inactive