EasyCache  by H-EmbodVis

Accelerate video diffusion inference without retraining

Created 4 months ago
259 stars

Top 98.0% on SourcePulse

GitHubView on GitHub
Project Summary

Summary

EasyCache addresses the slow inference speeds and high computational costs of video diffusion models, which impede their practical application. This training-free framework accelerates video generation by employing a runtime-adaptive caching mechanism to reuse computed transformation vectors, avoiding redundant computations. It targets researchers and developers needing efficient, high-quality video synthesis, offering substantial performance boosts and improved accessibility.

How It Works

The core of EasyCache is a lightweight, runtime-adaptive caching strategy that dynamically reuses previously computed transformation vectors during the iterative denoising process. This avoids redundant computations without requiring offline profiling, pre-computation, or extensive parameter tuning, offering a simple, immediately applicable solution for efficient video generation.

Quick Start & Requirements

Detailed usage instructions for each supported model are in their respective directories. The project supports models like HunyuanVideo, Wan2.1, and Wan2.2. GPU acceleration is implicitly required, with performance benchmarks on NVIDIA A800/H20 GPUs. Further details are on the Project Homepage and arXiv paper.

Highlighted Details

  • Achieves significant inference speedups, reducing latency by up to 2.1-3.3x compared to baselines on models like HunyuanVideo and Wan2.1.
  • Demonstrates up to a 36% PSNR improvement over previous state-of-the-art training-free acceleration methods.
  • Is orthogonal to other acceleration techniques, such as SVG attention, allowing combined performance gains (e.g., 3.33x speedup with SVG on HunyuanVideo).
  • Offers ComfyUI integration via community wrappers for various models.

Maintenance & Community

Recent releases in late July 2025 expanded support to multiple Wan2.2 models. Community contributions include ComfyUI wrappers for integration. No direct community channels (e.g., Discord, Slack) are listed.

Licensing & Compatibility

The code is licensed under the permissive Apache 2.0 License, generally compatible with commercial use and closed-source applications.

Limitations & Caveats

The README does not explicitly detail limitations. However, as a developing project with expanding model support, users may encounter initial limitations in compatibility breadth or edge cases. The focus on specific large-scale video generation models suggests general applicability to all diffusion models may require further development.

Health Check
Last Commit

2 months ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
2
Star History
9 stars in the last 30 days

Explore Similar Projects

Starred by Yineng Zhang Yineng Zhang(Inference Lead at SGLang; Research Scientist at Together AI), Yaowei Zheng Yaowei Zheng(Author of LLaMA-Factory), and
1 more.

FastVideo by hao-ai-lab

1.2%
3k
Framework for accelerated video generation
Created 1 year ago
Updated 2 days ago
Feedback? Help us improve.