ComfyUI plugin for efficient 4-bit neural network inference
Top 24.7% on sourcepulse
This repository provides ComfyUI nodes for Nunchaku, an efficient inference engine for 4-bit neural networks quantized with SVDQuant. It targets users of ComfyUI looking to leverage highly optimized, memory-efficient diffusion models, offering significant speedups and reduced VRAM requirements.
How It Works
Nunchaku utilizes SVDQuant for 4-bit quantization, enabling efficient inference on consumer hardware. The ComfyUI nodes integrate this engine, providing specialized loaders for diffusion models, LoRAs, and text encoders. Key advantages include a custom FP16 attention implementation that outperforms flash-attention2 on compatible hardware and a First-Block Cache mechanism to further accelerate inference.
Quick Start & Requirements
ComfyUI/custom_nodes
.comfy-cli
(optional). Requires downloading specific models (e.g., FLUX.1-schnell, text encoders) from HuggingFace/ModelScope.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
19 hours ago
1 day