Local AI voice chat for real-time conversations
Top 51.9% on sourcepulse
This project provides a local, real-time AI voice chat experience using the Zephyr 7B model, targeting users who want a customizable, offline conversational AI. It enables interactive voice communication with an AI that has a user-defined personality and voice.
How It Works
The system integrates Zephyr 7B via llama_cpp
for natural language processing, faster_whisper
(via RealtimeSTT
) for fast speech-to-text, and Coqui XTTS (via RealtimeTTS
) for high-quality, real-time text-to-speech. This stack allows for low-latency, end-to-end voice interaction without relying on external cloud services.
Quick Start & Requirements
pip install RealtimeSTT==0.1.7 RealtimeTTS==0.2.7
and pip install llama-cpp-python --force-reinstall --upgrade --no-cache-dir --verbose
.zephyr-7b-beta.Q5_K_M.gguf
and configure creation_params.json
.Highlighted Details
.wav
files.Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The software is in an experimental alpha state and may lack production stability. The XTTS model can exhibit glitches, and Zephyr 7B's quality is not comparable to larger models like GPT-4. Dependency conflicts may require manual installation of specific library versions.
1 month ago
1 day