Discover and explore top open-source AI tools and projects—updated daily.
Rust transcription tool using OpenAI's Whisper model
Top 84.3% on SourcePulse
Whisper Burn offers a Rust implementation of OpenAI's Whisper speech-to-text model, targeting developers seeking native Rust performance and control. It leverages the Burn deep learning framework, enabling efficient inference without Python dependencies.
How It Works
This project translates the Whisper architecture into Rust, utilizing Burn's tensor operations and backend flexibility (e.g., tch
for LibTorch, wgpu
for GPU acceleration). This approach aims for lower overhead and potentially faster execution compared to Python-based solutions, especially in resource-constrained environments or when integrating into existing Rust applications.
Quick Start & Requirements
git clone https://github.com/Gadersd/whisper-burn.git
).wget
, sox
(for audio resampling). For wgpu
backend, a compatible GPU and drivers are needed.tiny_en
).cargo run --release --bin transcribe <model_name> <audio_file> <language> <output_file>
. Example: cargo run --release --bin transcribe tiny_en audio16k.wav en transcription.txt
.sox audio.wav -r 16000 -c 1 audio16k.wav
.Highlighted Details
tch
, wgpu
).Maintenance & Community
The project is maintained by Gadersd. Community channels are not explicitly mentioned in the README.
Licensing & Compatibility
Limitations & Caveats
The wgpu
backend is noted as potentially unstable for large models. Conversion scripts require tinygrad
installed from source, which may add complexity to the setup.
1 year ago
1 day