Swift SDK for audio transcription
Top 49.2% on sourcepulse
This project provides a Swift package for easy integration of OpenAI's Whisper speech-to-text model into iOS, macOS, and other Apple platforms. It targets Swift developers seeking to add robust transcription capabilities to their applications, leveraging the performance of whisper.cpp.
How It Works
SwiftWhisper acts as a Swift wrapper around the whisper.cpp library, enabling direct use of Whisper models within Swift projects. It supports loading Whisper models and processing audio data, with optional Core ML acceleration for enhanced performance on Apple hardware. The library provides delegate methods for real-time progress updates and transcribed segments.
Quick Start & Requirements
https://github.com/exPHAT/SwiftWhisper.git
to your Package.swift
or Xcode's Package Manager tab.-encoder.mlmodelc
file alongside the main model file.fast
branch via a specific revision.Highlighted Details
Maintenance & Community
The project is hosted on GitHub. No specific community channels or active maintainer information are detailed in the README.
Licensing & Compatibility
The project appears to be MIT licensed, allowing for commercial use and integration into closed-source applications.
Limitations & Caveats
Transcription performance is significantly impacted by the build configuration (Debug vs. Release). Core ML support requires specific model file naming and placement. Audio conversion to the required PCM format is an additional step for developers.
1 year ago
1 day