whisper_ggml 1.0.0
whisper_ggml: ^1.0.0 copied to clipboard
OpenAI Whisper ASR (Automatic Speech Recognition) for Flutter
Whisper GGML #
OpenAI Whisper ASR (Automatic Speech Recognition) for Flutter using Whisper.cpp.
This project is a custom implementation of Whisper C++.
Features #
- Automatic Speech Recognition integration for Flutter apps.
- Supports automatic model downloading and initialization. Can be configured to work fully offline with models from
assetsfolder. - Seamless iOS and Android support with optimized performance.
- Utilizes CORE ML for enhanced processing on iOS devices.
Installation #
To use this library in your Flutter project, follow these steps:
- Add the library to your Flutter project's
pubspec.yaml:dependencies: whisper_ggml: ^1.0.0 - Run
flutter pub getto install the package.
Usage #
To integrate Whisper ASR in your Flutter app:
-
Import the package:
import 'package:whisper_ggml/whisper_ggml.dart'; -
Pick your model. Smaller models are more performant, but the accuracy may be lower. Recommended models are
tinyandsmall.final model = WhisperModel.tiny; -
Declare
WhisperControllerand use it for transcription:final controller = WhisperController(); final result = await whisperController.transcribe( model: model, /// Selected WhisperModel audioPath: audioPath, /// Path to .wav file lang: 'en', /// Language to transcribe ); -
Use the
resultvariable to access the transcription result:if (result?.transcription.text != null) { /// Do something with the transcription print(result!.transcription.text); }}
Notes #
Transcription processing time is about 5x times faster when running in release mode.