Discover and explore top open-source AI tools and projects—updated daily.
Interactive digital human conversation on a single PC
Top 18.9% on SourcePulse
This project provides a modular, interactive digital human conversation system designed to run on a single PC, targeting developers and researchers in AI and virtual reality. It offers low-latency, multimodal conversations with customizable components, enabling flexible integration of various AI models for speech, language, and avatar rendering.
How It Works
The system employs a modular architecture, allowing users to swap components for Automatic Speech Recognition (ASR), Large Language Models (LLM), Text-to-Speech (TTS), and avatar rendering. It supports both a fully local mode using models like MiniCPM-o and a hybrid mode leveraging cloud APIs for LLM and TTS. This flexibility reduces system requirements and allows for diverse conversational experiences.
Quick Start & Requirements
uv
for environment management. Install dependencies via uv sync --all-packages
or mode-specific installs. Run via uv run src/demo.py --config <config_file.yaml>
. Docker execution is also supported via ./build_and_run.sh --config <config_file.yaml>
.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
pynini
compilation issues.2 days ago
1 week