Talking face animation from a single image and audio
Top 3.9% on sourcepulse
SadTalker generates talking head videos from a single portrait image and an audio file. It is designed for researchers and users interested in AI-driven animation, offering a method to create stylized, realistic facial movements synchronized with speech. The project provides a user-friendly interface and multiple integration options.
How It Works
SadTalker leverages a 3D motion coefficient learning approach. It maps audio features to 3D facial motion parameters, which are then used to animate a single input portrait. This method allows for the generation of realistic head movements and expressions driven by the provided audio, with an optional enhancer (GFPGAN) for improved visual quality.
Quick Start & Requirements
requirements.txt
. PyTorch 1.12.1 with CUDA 11.3 is recommended.app_sadtalker.py
or webui.sh
/webui.bat
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project's disclaimer states it is not an official Tencent product and users must comply with applicable laws and intellectual property rights. It prohibits use for harmful activities or violations of social ethics.
1 year ago
Inactive