Discover and explore top open-source AI tools and projects—updated daily.
JasonzztGenAI DiT model acceleration for ComfyUI
Top 93.4% on SourcePulse
ComfyUI-CacheDiT provides a one-click solution to accelerate Diffusion Transformer (DiT) models within the ComfyUI environment. It targets users of ComfyUI who leverage DiT architectures for image and video generation, offering significant speedups (1.4-2.0x) with minimal configuration and no perceivable quality degradation. The primary benefit is reducing inference times for computationally intensive DiT models.
How It Works
The node implements an intelligent caching strategy inspired by llm-scaler. After an initial warmup phase, it selectively reuses previously computed intermediate results based on a skip_interval and the current inference step. If the conditions are met, cached data is utilized; otherwise, new computations are performed, and the result is cached for future steps. This approach minimizes redundant computations, leading to substantial performance gains.
Quick Start & Requirements
ComfyUI/custom_nodes/ and install dependencies via pip install -r requirements.txt.Highlighted Details
LTX2 Cache Optimizer, Wan Cache Optimizer) for specialized architectures like LTX-2 and WAN2.2 14B to ensure optimal performance and temporal consistency.Maintenance & Community
No specific details regarding maintainers, sponsorships, partnerships, or community channels (like Discord/Slack) are provided in the README.
Licensing & Compatibility
Limitations & Caveats
The speedup benefit is significantly reduced for inference tasks with very low step counts (less than 6 steps) due to warmup overhead. Model auto-detection may occasionally fail, requiring manual selection of the model_type preset. A 0% cache hit rate can occur if the model is not detected, inference steps are too short, or specific log messages are absent. Support for distilled low-step models beyond Z-Image-Turbo requires further validation.
2 months ago
Inactive
vllm-project