Discover and explore top open-source AI tools and projects—updated daily.
altalt-orgAccelerated local speech transcription for Apple Silicon
Top 77.3% on SourcePulse
Summary Lightning-SimulWhisper delivers high-performance, real-time local speech-to-text transcription optimized for Apple Silicon devices. It addresses the demand for efficient on-device processing by integrating MLX and CoreML, yielding substantial speed gains and enhanced power efficiency over standard PyTorch solutions. This project benefits users requiring rapid, responsive transcription without cloud dependencies.
How It Works The project employs a hybrid architecture for Whisper transcription on Apple Silicon. CoreML accelerates encoder inference via the Apple Neural Engine, achieving up to 18x speedups and reduced power draw. MLX manages the decoder, offering up to 15x speed improvements over PyTorch and supporting the AlignAtt policy for simultaneous decoding and configurable beam search. This dual-framework approach optimizes both transcription speed and system efficiency.
Quick Start & Requirements
pip install -r requirements.txt. For CoreML acceleration: pip install coremltools ane_transformers.whisper.cpp and generating CoreML encoder models via ./scripts/generate_coreml_encoder.sh <model_name>.Highlighted Details
tiny to large-v3-turbo.Maintenance & Community Details regarding maintainers, community channels (e.g., Discord, Slack), or a public roadmap are not provided in the README.
Licensing & Compatibility The specific open-source license governing this project is not explicitly stated in the provided README. Clarification is needed regarding commercial use or closed-source linking compatibility.
Limitations & Caveats
whisper.cpp) and executing specific scripts.2 months ago
Inactive
kyutai-labs
Vaibhavs10