Faster Whisper reimplementation using CTranslate2
Top 2.7% on sourcepulse
This project provides a significantly faster and more memory-efficient implementation of OpenAI's Whisper speech-to-text model, leveraging the CTranslate2 inference engine. It targets developers and researchers needing high-throughput transcription, offering up to 4x speed improvements and reduced resource consumption, especially with 8-bit quantization.
How It Works
Faster-Whisper reimplements the Whisper architecture using CTranslate2, a specialized C++ inference engine optimized for Transformer models. This allows for efficient execution on both CPU and GPU, with particular benefits from 8-bit quantization, which drastically reduces memory usage and speeds up computation without significant accuracy loss.
Quick Start & Requirements
pip install faster-whisper
pip
on Linux is supported.Highlighted Details
openai/whisper
on GPU (FP16) and significantly faster on CPU (INT8).Maintenance & Community
Licensing & Compatibility
Limitations & Caveats
ctranslate2
.2 months ago
1 day