Real-time digital human for mobile and web
Top 26.8% on sourcepulse
This project provides a real-time live streaming digital human solution, primarily focusing on DH_live_mini, designed for broad accessibility across mobile and web platforms without requiring GPUs. It targets users seeking an efficient, easy-to-deploy digital human avatar for applications like live streaming and real-time conversations.
How It Works
DH_live_mini utilizes a highly optimized approach, achieving a low single-frame compute power of 39 Mflops, significantly less than most mobile face detection algorithms. This efficiency allows it to run directly in mobile browsers and on CPUs, with a web resource package compressable to under 3MB. The system supports real-time dialogue, integrating Voice Activity Detection (VAD), Automatic Speech Recognition (ASR), Large Language Models (LLM), Text-to-Speech (TTS), and the digital human rendering pipeline.
Quick Start & Requirements
conda create -n dh_live python=3.11
, conda activate dh_live
, pip install torch --index-url https://download.pytorch.org/whl/cu124
(or CPU version), pip install -r requirements.txt
.data_preparation_mini.py
and data_preparation_web.py
.python web_demo/server.py
to access localhost:8888/static/MiniLive.html
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The original DH_live is no longer supported. Offline video synthesis is not supported on Linux/macOS. Audio file processing for demo_mini.py
is not supported on Linux/macOS.
1 month ago
1 week