TaylorSeer  by Shenyi-Z

Accelerating diffusion models with predictive feature caching

Created 6 months ago
284 stars

Top 92.1% on SourcePulse

GitHubView on GitHub
Project Summary

TaylorSeer accelerates Diffusion Transformer (DiT) models for image and video synthesis by predicting future timestep features using Taylor series expansion, enabling significant speedups without retraining. It targets researchers and developers working with high-fidelity generative models who need to reduce inference latency for real-time applications.

How It Works

TaylorSeer leverages the observation that diffusion model features evolve slowly and continuously across timesteps. It approximates higher-order derivatives of these features to predict future states via Taylor series expansion. This forecasting approach aims to overcome the quality degradation seen in traditional feature caching methods when timestep intervals are large, offering substantial acceleration with minimal impact on generation quality.

Quick Start & Requirements

Highlighted Details

  • Achieves 4.99x lossless compression and 3.53x latency speedup on FLUX.1-dev.
  • Achieves 5.00x compression and 4.65x latency speedup on HunyuanVideo.
  • On DiT, achieves 3.41 lower FID than SOTA at 4.53x acceleration.
  • Supports multi-GPU parallel inference for models like HunyuanVideo and Wan2.1.

Maintenance & Community

The project is associated with ICCV 2025 and ICLR 2025 submissions. It acknowledges contributions from various model implementations (DiT, FLUX, HiDream, etc.) and has community contributions like ComfyUI-TaylorSeer. Contact email: shenyizou@outlook.com.

Licensing & Compatibility

The repository does not explicitly state a license. Compatibility for commercial use or closed-source linking is not specified.

Limitations & Caveats

The project is presented as research code, with specific implementations for various models. While claiming "lossless" or "near lossless" acceleration, the exact quality metrics and potential trade-offs at higher acceleration ratios may require further investigation. The primary focus is on DiT architectures and related models.

Health Check
Last Commit

1 month ago

Responsiveness

Inactive

Pull Requests (30d)
0
Issues (30d)
4
Star History
24 stars in the last 30 days

Explore Similar Projects

Starred by Alex Yu Alex Yu(Research Scientist at OpenAI; Former Cofounder of Luma AI) and Cody Yu Cody Yu(Coauthor of vLLM; MTS at OpenAI).

xDiT by xdit-project

0.7%
2k
Inference engine for parallel Diffusion Transformer (DiT) deployment
Created 1 year ago
Updated 1 day ago
Feedback? Help us improve.