AI-powered personality core for interactive, embodied assistant
Top 10.2% on sourcepulse
This project aims to create a physical, interactive AI embodying GLaDOS from the Portal series, targeting hobbyists and developers interested in embodied AI and robotics. It offers a unique opportunity to build a sophisticated conversational agent with a physical presence, capable of low-latency voice interaction and potential future vision capabilities.
How It Works
The system employs a low-latency pipeline: continuous audio recording buffers data, detecting voice activity. Upon cessation, speech is transcribed and streamed to a local LLM. Sentence-by-sentence LLM output is fed to a text-to-speech engine, enabling concurrent generation and playback for reduced latency. The architecture prioritizes minimal dependencies for constrained hardware, avoiding large frameworks like PyTorch.
Quick Start & Requirements
python scripts/install.py
(or scripts\install.py
on Windows).uv run glados
or uv run glados tui
for the Text UI.llama3.2
via Ollama) and an OpenAI-compatible TTS server. Performance is highly dependent on hardware acceleration.Highlighted Details
Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
The project is in active, experimental development, particularly the SBC implementation, and does not guarantee support for complex setup issues. Users may encounter segfaults and require significant troubleshooting, especially on non-standard hardware. Voice interruption loops can occur without proper audio hardware or configuration.
1 week ago
Inactive