whisper_ggml 1.0.0
whisper_ggml: ^1.0.0 copied to clipboard
OpenAI Whisper ASR (Automatic Speech Recognition) for Flutter
example/lib/main.dart
import 'dart:io';
import 'package:flutter/material.dart';
import 'package:flutter/services.dart';
import 'package:path_provider/path_provider.dart';
import 'package:whisper_ggml/whisper_ggml.dart';
import 'package:record/record.dart';
void main() {
runApp(const MyApp());
}
class MyApp extends StatelessWidget {
const MyApp({super.key});
@override
Widget build(BuildContext context) {
return MaterialApp(
title: 'Whisper ggml example',
theme: ThemeData(
colorScheme: ColorScheme.fromSeed(seedColor: Colors.deepPurple),
useMaterial3: true,
),
home: const MyHomePage(),
);
}
}
class MyHomePage extends StatefulWidget {
/// Modify this model based on your needs
const MyHomePage({super.key});
@override
State<MyHomePage> createState() => _MyHomePageState();
}
class _MyHomePageState extends State<MyHomePage> {
final model = WhisperModel.base;
final AudioRecorder audioRecorder = AudioRecorder();
final WhisperController whisperController = WhisperController();
String transcribedText = 'Transcribed text will be displayed here';
bool isProcessing = false;
bool isListening = false;
@override
void initState() {
initModel();
super.initState();
}
@override
Widget build(BuildContext context) {
return Scaffold(
appBar: AppBar(
backgroundColor: Theme.of(context).colorScheme.inversePrimary,
title: Text('Whisper ggml example'),
),
body: Center(
child: Padding(
padding: const EdgeInsets.symmetric(horizontal: 24),
child: Text(
transcribedText,
style: Theme.of(context).textTheme.headlineMedium,
),
),
),
floatingActionButton: FloatingActionButton(
onPressed: record,
tooltip: 'Start listening',
child: isProcessing
? const CircularProgressIndicator()
: Icon(
isListening ? Icons.mic_off : Icons.mic,
color: isListening ? Colors.red : null,
),
),
);
}
Future<void> initModel() async {
try {
/// Try initializing the model from assets
final bytesBase = await rootBundle.load('assets/ggml-${model.modelName}.bin');
final modelPathBase = await whisperController.getPath(model);
final fileBase = File(modelPathBase);
await fileBase.writeAsBytes(bytesBase.buffer.asUint8List(bytesBase.offsetInBytes, bytesBase.lengthInBytes));
} catch (e) {
/// On error try downloading the model
await whisperController.downloadModel(model);
}
}
Future<void> record() async {
if (await audioRecorder.hasPermission()) {
if (await audioRecorder.isRecording()) {
final audioPath = await audioRecorder.stop();
if (audioPath != null) {
debugPrint('Stopped listening.');
setState(() {
isListening = false;
isProcessing = true;
});
final result = await whisperController.transcribe(
model: model,
audioPath: audioPath,
lang: 'en',
);
if (mounted) {
setState(() {
isProcessing = false;
});
}
if (result?.transcription.text != null) {
setState(() {
transcribedText = result!.transcription.text;
});
}
} else {
debugPrint('No recording exists.');
}
} else {
debugPrint('Started listening.');
setState(() {
isListening = true;
});
final Directory appDirectory = await getTemporaryDirectory();
await audioRecorder.start(const RecordConfig(), path: '${appDirectory.path}/test.m4a');
}
}
}
}