Discover and explore top open-source AI tools and projects—updated daily.
linyiLYiVoice assistant demo using local LLM
Top 30.3% on SourcePulse
This project provides a local, voice-activated assistant for dialogue with large language models. It's a toy demo targeting users interested in offline AI interaction, offering a simple Python script for voice recognition and text generation.
How It Works
The assistant leverages Apple's MLX implementation of OpenAI's Whisper for voice-to-text conversion. For text generation, it utilizes the Yi large language model (e.g., Yi-34B-Chat or Yi-6B-Chat) running locally via LangChain and llama.cpp. This approach enables offline operation and direct interaction without relying on cloud APIs.
Quick Start & Requirements
conda create -n VoiceAI python=3.11, conda activate VoiceAI, pip install -r requirements.txt, CMAKE_ARGS="-DLLAMA_METAL=on" pip install llama-cpp-python.portaudio (via brew install portaudio), pyaudio. Tested on macOS; Windows/Linux require alternative audio libraries.models/. Whisper model (e.g., whisper-large-v3) also required in models/.Highlighted Details
llama.cpp.main.py) for simplicity.llama-cpp-python build flags.Maintenance & Community
The project is a personal toy demo. No specific community channels or roadmap are indicated.
Licensing & Compatibility
The README does not explicitly state a license. The project uses components from Apple MLX and OpenAI Whisper, which have their own licenses. Compatibility for commercial use or closed-source linking is not specified.
Limitations & Caveats
The project is described as a "toy demo" and has only been tested on macOS, with Windows/Linux requiring manual replacement of audio components. Model file paths are hardcoded variables, and the 34B model has significant hardware requirements.
1 year ago
Inactive