Discover and explore top open-source AI tools and projects—updated daily.
soniqoApple Silicon speech AI toolkit
Top 75.7% on SourcePulse
Summary
This project provides a comprehensive suite of AI speech models (ASR, TTS, speech-to-speech, VAD, diarization) optimized for on-device execution on Apple Silicon. It targets developers building native macOS and iOS applications, offering significant performance benefits and privacy by leveraging MLX and CoreML frameworks.
How It Works
The toolkit integrates advanced speech models, including Qwen3-ASR, CosyVoice TTS, and PersonaPlex, utilizing Apple's MLX library for GPU acceleration and CoreML for Neural Engine efficiency. This dual-backend approach allows for high-throughput processing or power-optimized inference, enabling real-time speech applications directly on user devices without cloud dependencies.
Quick Start & Requirements
Installation is streamlined via Homebrew (brew install speech) or Swift Package Manager. Building from source requires git clone and make build. Essential prerequisites include Swift 5.9+, macOS 14+ or iOS 17+, Apple Silicon hardware, and Xcode 15+ with the Metal Toolchain. Initial model downloads can range from megabytes to several gigabytes.
Highlighted Details
PersonaPlexDemo for a conversational voice assistant.Maintenance & Community
The project actively incorporates recent advancements, with recent news highlighting new features like speaker diarization and PersonaPlex integration. A roadmap discussion is available for community input, and contributions via pull requests are welcomed.
Licensing & Compatibility
The project is released under the permissive Apache 2.0 license, ensuring broad compatibility for commercial use and integration into closed-source applications.
Limitations & Caveats
Strictly limited to Apple Silicon hardware and recent macOS/iOS versions; Rosetta/x86_64 architectures are unsupported. Successful GPU acceleration via MLX depends on correctly building the MLX Metal library, which can be a setup hurdle. CoreML offers power efficiency but may yield lower throughput for single-model tasks compared to MLX.
21 hours ago
Inactive
moonshine-ai