Add TTS API and examples for Dart (#1010)
This commit is contained in:
69
dart-api-examples/tts/bin/coqui.dart
Normal file
69
dart-api-examples/tts/bin/coqui.dart
Normal file
@@ -0,0 +1,69 @@
|
||||
// Copyright (c) 2024 Xiaomi Corporation
|
||||
import 'dart:io';
|
||||
import 'dart:typed_data';
|
||||
|
||||
import 'package:args/args.dart';
|
||||
import 'package:sherpa_onnx/sherpa_onnx.dart' as sherpa_onnx;
|
||||
|
||||
import './init.dart';
|
||||
|
||||
void main(List<String> arguments) async {
|
||||
await initSherpaOnnx();
|
||||
|
||||
final parser = ArgParser()
|
||||
..addOption('model', help: 'Path to the ONNX model')
|
||||
..addOption('tokens', help: 'Path to tokens.txt')
|
||||
..addOption('text', help: 'Text to generate TTS for')
|
||||
..addOption('output-wav', help: 'Filename to save the generated audio')
|
||||
..addOption('speed', help: 'Speech speed', defaultsTo: '1.0')
|
||||
..addOption(
|
||||
'sid',
|
||||
help: 'Speaker ID to select. Used only for multi-speaker TTS',
|
||||
defaultsTo: '0',
|
||||
);
|
||||
final res = parser.parse(arguments);
|
||||
if (res['model'] == null ||
|
||||
res['tokens'] == null ||
|
||||
res['output-wav'] == null ||
|
||||
res['text'] == null) {
|
||||
print(parser.usage);
|
||||
exit(1);
|
||||
}
|
||||
final model = res['model'] as String;
|
||||
final tokens = res['tokens'] as String;
|
||||
final text = res['text'] as String;
|
||||
final outputWav = res['output-wav'] as String;
|
||||
var speed = double.tryParse(res['speed'] as String) ?? 1.0;
|
||||
final sid = int.tryParse(res['sid'] as String) ?? 0;
|
||||
|
||||
if (speed == 0) {
|
||||
speed = 1.0;
|
||||
}
|
||||
|
||||
final vits = sherpa_onnx.OfflineTtsVitsModelConfig(
|
||||
model: model,
|
||||
tokens: tokens,
|
||||
lengthScale: 1 / speed,
|
||||
);
|
||||
|
||||
final modelConfig = sherpa_onnx.OfflineTtsModelConfig(
|
||||
vits: vits,
|
||||
numThreads: 1,
|
||||
debug: true,
|
||||
);
|
||||
final config = sherpa_onnx.OfflineTtsConfig(
|
||||
model: modelConfig,
|
||||
maxNumSenetences: 1,
|
||||
);
|
||||
|
||||
final tts = sherpa_onnx.OfflineTts(config);
|
||||
final audio = tts.generate(text: text, sid: sid, speed: speed);
|
||||
tts.free();
|
||||
|
||||
sherpa_onnx.writeWave(
|
||||
filename: outputWav,
|
||||
samples: audio.samples,
|
||||
sampleRate: audio.sampleRate,
|
||||
);
|
||||
print('Saved to ${outputWav}');
|
||||
}
|
||||
1
dart-api-examples/tts/bin/init.dart
Symbolic link
1
dart-api-examples/tts/bin/init.dart
Symbolic link
@@ -0,0 +1 @@
|
||||
../../vad/bin/init.dart
|
||||
80
dart-api-examples/tts/bin/piper.dart
Normal file
80
dart-api-examples/tts/bin/piper.dart
Normal file
@@ -0,0 +1,80 @@
|
||||
// Copyright (c) 2024 Xiaomi Corporation
|
||||
import 'dart:io';
|
||||
import 'dart:typed_data';
|
||||
|
||||
import 'package:args/args.dart';
|
||||
import 'package:sherpa_onnx/sherpa_onnx.dart' as sherpa_onnx;
|
||||
|
||||
import './init.dart';
|
||||
|
||||
void main(List<String> arguments) async {
|
||||
await initSherpaOnnx();
|
||||
|
||||
final parser = ArgParser()
|
||||
..addOption('model', help: 'Path to the ONNX model')
|
||||
..addOption('tokens', help: 'Path to tokens.txt')
|
||||
..addOption('data-dir', help: 'Path to espeak-ng-data directory')
|
||||
..addOption('text', help: 'Text to generate TTS for')
|
||||
..addOption('output-wav', help: 'Filename to save the generated audio')
|
||||
..addOption('speed', help: 'Speech speed', defaultsTo: '1.0')
|
||||
..addOption(
|
||||
'sid',
|
||||
help: 'Speaker ID to select. Used only for multi-speaker TTS',
|
||||
defaultsTo: '0',
|
||||
);
|
||||
final res = parser.parse(arguments);
|
||||
if (res['model'] == null ||
|
||||
res['tokens'] == null ||
|
||||
res['data-dir'] == null ||
|
||||
res['output-wav'] == null ||
|
||||
res['text'] == null) {
|
||||
print(parser.usage);
|
||||
exit(1);
|
||||
}
|
||||
final model = res['model'] as String;
|
||||
final tokens = res['tokens'] as String;
|
||||
final dataDir = res['data-dir'] as String;
|
||||
final text = res['text'] as String;
|
||||
final outputWav = res['output-wav'] as String;
|
||||
var speed = double.tryParse(res['speed'] as String) ?? 1.0;
|
||||
final sid = int.tryParse(res['sid'] as String) ?? 0;
|
||||
|
||||
if (speed == 0) {
|
||||
speed = 1.0;
|
||||
}
|
||||
|
||||
final vits = sherpa_onnx.OfflineTtsVitsModelConfig(
|
||||
model: model,
|
||||
tokens: tokens,
|
||||
dataDir: dataDir,
|
||||
lengthScale: 1 / speed,
|
||||
);
|
||||
|
||||
final modelConfig = sherpa_onnx.OfflineTtsModelConfig(
|
||||
vits: vits,
|
||||
numThreads: 1,
|
||||
debug: true,
|
||||
);
|
||||
final config = sherpa_onnx.OfflineTtsConfig(
|
||||
model: modelConfig,
|
||||
maxNumSenetences: 1,
|
||||
);
|
||||
|
||||
final tts = sherpa_onnx.OfflineTts(config);
|
||||
final audio = tts.generateWithCallback(
|
||||
text: text,
|
||||
sid: sid,
|
||||
speed: speed,
|
||||
callback: (Float32List samples) {
|
||||
print('${samples.length} samples received');
|
||||
// You can play samples in a separate thread/isolate
|
||||
});
|
||||
tts.free();
|
||||
|
||||
sherpa_onnx.writeWave(
|
||||
filename: outputWav,
|
||||
samples: audio.samples,
|
||||
sampleRate: audio.sampleRate,
|
||||
);
|
||||
print('Saved to ${outputWav}');
|
||||
}
|
||||
86
dart-api-examples/tts/bin/zh.dart
Normal file
86
dart-api-examples/tts/bin/zh.dart
Normal file
@@ -0,0 +1,86 @@
|
||||
// Copyright (c) 2024 Xiaomi Corporation
|
||||
import 'dart:io';
|
||||
import 'dart:typed_data';
|
||||
|
||||
import 'package:args/args.dart';
|
||||
import 'package:sherpa_onnx/sherpa_onnx.dart' as sherpa_onnx;
|
||||
|
||||
import './init.dart';
|
||||
|
||||
void main(List<String> arguments) async {
|
||||
await initSherpaOnnx();
|
||||
|
||||
final parser = ArgParser()
|
||||
..addOption('model', help: 'Path to the ONNX model')
|
||||
..addOption('tokens', help: 'Path to tokens.txt')
|
||||
..addOption('lexicon', help: 'Path to lexicon.txt')
|
||||
..addOption(
|
||||
'dict-dir',
|
||||
help: 'Path to jieba dict directory',
|
||||
defaultsTo: '',
|
||||
)
|
||||
..addOption('rule-fsts', help: 'Path to rule fsts', defaultsTo: '')
|
||||
..addOption('rule-fars', help: 'Path to rule fars', defaultsTo: '')
|
||||
..addOption('text', help: 'Text to generate TTS for')
|
||||
..addOption('output-wav', help: 'Filename to save the generated audio')
|
||||
..addOption('speed', help: 'Speech speed', defaultsTo: '1.0')
|
||||
..addOption(
|
||||
'sid',
|
||||
help: 'Speaker ID to select. Used only for multi-speaker TTS',
|
||||
defaultsTo: '0',
|
||||
);
|
||||
final res = parser.parse(arguments);
|
||||
if (res['model'] == null ||
|
||||
res['lexicon'] == null ||
|
||||
res['tokens'] == null ||
|
||||
res['output-wav'] == null ||
|
||||
res['text'] == null) {
|
||||
print(parser.usage);
|
||||
exit(1);
|
||||
}
|
||||
final model = res['model'] as String;
|
||||
final lexicon = res['lexicon'] as String;
|
||||
final tokens = res['tokens'] as String;
|
||||
final dictDir = res['dict-dir'] as String;
|
||||
final ruleFsts = res['rule-fsts'] as String;
|
||||
final ruleFars = res['rule-fars'] as String;
|
||||
final text = res['text'] as String;
|
||||
final outputWav = res['output-wav'] as String;
|
||||
var speed = double.tryParse(res['speed'] as String) ?? 1.0;
|
||||
final sid = int.tryParse(res['sid'] as String) ?? 0;
|
||||
|
||||
if (speed == 0) {
|
||||
speed = 1.0;
|
||||
}
|
||||
|
||||
final vits = sherpa_onnx.OfflineTtsVitsModelConfig(
|
||||
model: model,
|
||||
lexicon: lexicon,
|
||||
tokens: tokens,
|
||||
dictDir: dictDir,
|
||||
lengthScale: 1 / speed,
|
||||
);
|
||||
|
||||
final modelConfig = sherpa_onnx.OfflineTtsModelConfig(
|
||||
vits: vits,
|
||||
numThreads: 1,
|
||||
debug: true,
|
||||
);
|
||||
final config = sherpa_onnx.OfflineTtsConfig(
|
||||
model: modelConfig,
|
||||
maxNumSenetences: 1,
|
||||
ruleFsts: ruleFsts,
|
||||
ruleFars: ruleFars,
|
||||
);
|
||||
|
||||
final tts = sherpa_onnx.OfflineTts(config);
|
||||
final audio = tts.generate(text: text, sid: sid, speed: speed);
|
||||
tts.free();
|
||||
|
||||
sherpa_onnx.writeWave(
|
||||
filename: outputWav,
|
||||
samples: audio.samples,
|
||||
sampleRate: audio.sampleRate,
|
||||
);
|
||||
print('Saved to ${outputWav}');
|
||||
}
|
||||
Reference in New Issue
Block a user