ComfyUI nodes for deep learning-based face manipulation
Top 71.2% on sourcepulse
DeepFuze is a ComfyUI custom node suite for advanced facial transformation, including lipsyncing, face swapping, and voice cloning. It targets content creators, animators, and developers seeking to enhance video projects with AI-driven realism and synchronization, offering a powerful offline solution.
How It Works
DeepFuze integrates advanced deep learning models within the ComfyUI node-based workflow. It leverages specialized nodes for tasks like lipsync generation, face swapping, and voice cloning, allowing users to combine audio and video with precise facial movement synchronization. The architecture supports various face detection models (YOLOFace, RetinaFace, SCRFD, YuNet) and includes optional enhancers for improved output quality.
Quick Start & Requirements
https://github.com/SamKhoze/ComfyUI-DeepFuze.git
.export PYTORCH_ENABLE_MPS_FALLBACK=1
environment variable. onnxruntime
(CPU version), dlib
, and TTS
must be installed.Highlighted Details
Maintenance & Community
The project is developed by Dr. Sam Khoze and his team. Links to community channels or roadmaps are not explicitly provided in the README.
Licensing & Compatibility
The code is released under an unspecified open-source license. It is stated to be free for personal, research, academic, and commercial use, with a caution to comply with applicable laws and use responsibly.
Limitations & Caveats
CUDA installation on Windows is noted as non-beginner-friendly. macOS installation requires manual steps outside the ComfyUI Manager. The TTS node may have sample rate issues with certain audio formats, with a converter node in development. The OpenAI LLM node requires manual API key entry on each use unless set as an environment variable.
8 months ago
1 week