Voice assistant for experimenting with SOTA voice models
Top 36.7% on sourcepulse
Verbi is a modular voice assistant framework designed for researchers and developers to experiment with and compare state-of-the-art (SOTA) speech and language models. It offers flexibility in swapping components for transcription, response generation, and text-to-speech (TTS), enabling easy evaluation of different AI services and local models.
How It Works
Verbi employs a modular architecture, allowing users to configure different SOTA models via a central config.py
file. It supports integrations with cloud APIs like OpenAI, Groq, and Deepgram, as well as local models through Ollama and dedicated local TTS servers (MeloTTS, Piper). This design facilitates rapid prototyping and comparative analysis of voice assistant technologies.
Quick Start & Requirements
pip install -r requirements.txt
.env
, configure config.py
.python run_voice_assistant.py
Highlighted Details
Maintenance & Community
The project is actively maintained by PromtEngineer. Community contributions are welcomed via pull requests.
Licensing & Compatibility
The repository's license is not explicitly stated in the README. Compatibility for commercial use or closed-source linking would require clarification of the licensing terms.
Limitations & Caveats
The README does not specify the project's license, which is crucial for determining commercial use and compatibility. Some local TTS models (MeloTTS, Piper) require separate installation and server setup.
2 months ago
1 day