Add speaker identification and verification exmaple for Dart API (#1194)
This commit is contained in:
5
.github/scripts/test-dart.sh
vendored
5
.github/scripts/test-dart.sh
vendored
@@ -4,6 +4,11 @@ set -ex
|
||||
|
||||
cd dart-api-examples
|
||||
|
||||
pushd speaker-identification
|
||||
echo '----------3d speaker----------'
|
||||
./run-3d-speaker.sh
|
||||
popd
|
||||
|
||||
pushd add-punctuations
|
||||
echo '----------CT Transformer----------'
|
||||
./run-ct-transformer.sh
|
||||
|
||||
1
.github/workflows/test-dart.yaml
vendored
1
.github/workflows/test-dart.yaml
vendored
@@ -112,6 +112,7 @@ jobs:
|
||||
cp scripts/dart/vad-non-streaming-asr-pubspec.yaml dart-api-examples/vad-with-non-streaming-asr/pubspec.yaml
|
||||
cp scripts/dart/audio-tagging-pubspec.yaml dart-api-examples/audio-tagging/pubspec.yaml
|
||||
cp scripts/dart/add-punctuations-pubspec.yaml dart-api-examples/add-punctuations/pubspec.yaml
|
||||
cp scripts/dart/speaker-id-pubspec.yaml dart-api-examples/speaker-identification/pubspec.yaml
|
||||
|
||||
cp scripts/dart/sherpa-onnx-pubspec.yaml flutter/sherpa_onnx/pubspec.yaml
|
||||
|
||||
|
||||
@@ -9,14 +9,15 @@ https://pub.dev/packages/sherpa_onnx
|
||||
|
||||
| Directory | Description |
|
||||
|-----------|-------------|
|
||||
| [./add-punctuations](./add-punctuations)| Example for adding punctuations to text.|
|
||||
| [./audio-tagging](./audio-tagging)| Example for audio tagging.|
|
||||
| [./keyword-spotter](./keyword-spotter)| Example for keyword spotting|
|
||||
| [./non-streaming-asr](./non-streaming-asr)| Example for non-streaming speech recognition|
|
||||
| [./speaker-identification](./speaker-identification)| Example for speaker identification and verification.|
|
||||
| [./streaming-asr](./streaming-asr)| Example for streaming speech recognition|
|
||||
| [./tts](./tts)| Example for text to speech|
|
||||
| [./vad](./vad)| Example for voice activity detection|
|
||||
| [./vad-with-non-streaming-asr](./vad-with-non-streaming-asr)| Example for voice activity detection with non-streaming speech recognition. You can use it to generate subtitles.|
|
||||
| [./audio-tagging](./audio-tagging)| Example for audio tagging.|
|
||||
| [./add-punctuations](./add-punctuations)| Example for adding punctuations to text.|
|
||||
| [./vad](./vad)| Example for voice activity detection|
|
||||
|
||||
## How to create an example in this folder
|
||||
|
||||
|
||||
3
dart-api-examples/speaker-identification/.gitignore
vendored
Normal file
3
dart-api-examples/speaker-identification/.gitignore
vendored
Normal file
@@ -0,0 +1,3 @@
|
||||
# https://dart.dev/guides/libraries/private-files
|
||||
# Created by `dart pub`
|
||||
.dart_tool/
|
||||
7
dart-api-examples/speaker-identification/README.md
Normal file
7
dart-api-examples/speaker-identification/README.md
Normal file
@@ -0,0 +1,7 @@
|
||||
# Introduction
|
||||
|
||||
This example shows how to use the Dart API from sherpa-onnx for speaker identification.
|
||||
|
||||
| File | Description|
|
||||
|------|------------|
|
||||
|[./bin/speaker_id.dart](./bin/speaker_id.dart)| Use a speaker embedding extractor model for speaker identification and verification. See also [./run-3d-speaker.sh](./run-3d-speaker.sh)|
|
||||
@@ -0,0 +1,30 @@
|
||||
# This file configures the static analysis results for your project (errors,
|
||||
# warnings, and lints).
|
||||
#
|
||||
# This enables the 'recommended' set of lints from `package:lints`.
|
||||
# This set helps identify many issues that may lead to problems when running
|
||||
# or consuming Dart code, and enforces writing Dart using a single, idiomatic
|
||||
# style and format.
|
||||
#
|
||||
# If you want a smaller set of lints you can change this to specify
|
||||
# 'package:lints/core.yaml'. These are just the most critical lints
|
||||
# (the recommended set includes the core lints).
|
||||
# The core lints are also what is used by pub.dev for scoring packages.
|
||||
|
||||
include: package:lints/recommended.yaml
|
||||
|
||||
# Uncomment the following section to specify additional rules.
|
||||
|
||||
# linter:
|
||||
# rules:
|
||||
# - camel_case_types
|
||||
|
||||
# analyzer:
|
||||
# exclude:
|
||||
# - path/to/excluded/files/**
|
||||
|
||||
# For more information about the core and recommended set of lints, see
|
||||
# https://dart.dev/go/core-lints
|
||||
|
||||
# For additional information about configuring this file, see
|
||||
# https://dart.dev/guides/language/analysis-options
|
||||
1
dart-api-examples/speaker-identification/bin/init.dart
Symbolic link
1
dart-api-examples/speaker-identification/bin/init.dart
Symbolic link
@@ -0,0 +1 @@
|
||||
../../vad/bin/init.dart
|
||||
160
dart-api-examples/speaker-identification/bin/speaker_id.dart
Normal file
160
dart-api-examples/speaker-identification/bin/speaker_id.dart
Normal file
@@ -0,0 +1,160 @@
|
||||
// Copyright (c) 2024 Xiaomi Corporation
|
||||
import 'dart:io';
|
||||
import 'dart:typed_data';
|
||||
|
||||
import 'package:args/args.dart';
|
||||
import 'package:sherpa_onnx/sherpa_onnx.dart' as sherpa_onnx;
|
||||
import './init.dart';
|
||||
|
||||
Float32List computeEmbedding(
|
||||
{required sherpa_onnx.SpeakerEmbeddingExtractor extractor,
|
||||
required String filename}) {
|
||||
final waveData = sherpa_onnx.readWave(filename);
|
||||
final stream = extractor.createStream();
|
||||
|
||||
stream.acceptWaveform(
|
||||
samples: waveData.samples,
|
||||
sampleRate: waveData.sampleRate,
|
||||
);
|
||||
|
||||
stream.inputFinished();
|
||||
|
||||
final embedding = extractor.compute(stream);
|
||||
|
||||
stream.free();
|
||||
|
||||
return embedding;
|
||||
}
|
||||
|
||||
void main(List<String> arguments) async {
|
||||
await initSherpaOnnx();
|
||||
|
||||
final parser = ArgParser()..addOption('model', help: 'Path to model.onnx');
|
||||
|
||||
final res = parser.parse(arguments);
|
||||
if (res['model'] == null) {
|
||||
print(parser.usage);
|
||||
exit(1);
|
||||
}
|
||||
|
||||
final model = res['model'] as String;
|
||||
/*
|
||||
Please download test data by yourself
|
||||
|
||||
curl -SL -o sr-data.tar.gz https://github.com/csukuangfj/sr-data/archive/refs/tags/v1.0.0.tar.gz
|
||||
tar xvf sr-data.tar.gz
|
||||
mv sr-data-1.0.0 sr-data
|
||||
*/
|
||||
|
||||
final config = sherpa_onnx.SpeakerEmbeddingExtractorConfig(
|
||||
model: model,
|
||||
numThreads: 1,
|
||||
debug: true,
|
||||
provider: 'cpu',
|
||||
);
|
||||
final extractor = sherpa_onnx.SpeakerEmbeddingExtractor(config: config);
|
||||
|
||||
final manager = sherpa_onnx.SpeakerEmbeddingManager(extractor.dim);
|
||||
|
||||
final spk1Files = [
|
||||
"./sr-data/enroll/fangjun-sr-1.wav",
|
||||
"./sr-data/enroll/fangjun-sr-2.wav",
|
||||
"./sr-data/enroll/fangjun-sr-3.wav",
|
||||
];
|
||||
|
||||
final spk1Vec = <Float32List>[];
|
||||
for (final f in spk1Files) {
|
||||
final embedding = computeEmbedding(extractor: extractor, filename: f);
|
||||
spk1Vec.add(embedding);
|
||||
}
|
||||
|
||||
final spk2Files = [
|
||||
"./sr-data/enroll/leijun-sr-1.wav",
|
||||
"./sr-data/enroll/leijun-sr-2.wav",
|
||||
];
|
||||
|
||||
final spk2Vec = <Float32List>[];
|
||||
for (final f in spk2Files) {
|
||||
final embedding = computeEmbedding(extractor: extractor, filename: f);
|
||||
spk2Vec.add(embedding);
|
||||
}
|
||||
|
||||
if (!manager.addMulti(name: "fangjun", embeddingList: spk1Vec)) {
|
||||
// Note you should free extractor and manager in your app to avoid memory leak
|
||||
print("Failed to register fangjun");
|
||||
return;
|
||||
}
|
||||
|
||||
if (!manager.addMulti(name: "leijun", embeddingList: spk2Vec)) {
|
||||
print("Failed to register leijun");
|
||||
return;
|
||||
}
|
||||
|
||||
if (manager.numSpeakers != 2) {
|
||||
print("There should be two speakers");
|
||||
return;
|
||||
}
|
||||
|
||||
if (!manager.contains("fangjun")) {
|
||||
print("It should contain the speaker fangjun");
|
||||
return;
|
||||
}
|
||||
|
||||
if (!manager.contains("leijun")) {
|
||||
print("It should contain the speaker leijun");
|
||||
return;
|
||||
}
|
||||
|
||||
print("---All speakers---");
|
||||
final allSpeakers = manager.allSpeakerNames;
|
||||
for (final s in allSpeakers) {
|
||||
print(s);
|
||||
}
|
||||
print("------------");
|
||||
|
||||
final testFiles = [
|
||||
"./sr-data/test/fangjun-test-sr-1.wav",
|
||||
"./sr-data/test/leijun-test-sr-1.wav",
|
||||
"./sr-data/test/liudehua-test-sr-1.wav",
|
||||
];
|
||||
|
||||
final threshold = 0.6;
|
||||
for (final file in testFiles) {
|
||||
final embedding = computeEmbedding(extractor: extractor, filename: file);
|
||||
|
||||
var name = manager.search(embedding: embedding, threshold: threshold);
|
||||
if (name == '') {
|
||||
name = "<Unknown>";
|
||||
}
|
||||
print("$file: $name");
|
||||
}
|
||||
|
||||
if (!manager.verify(
|
||||
name: "fangjun",
|
||||
embedding: computeEmbedding(extractor: extractor, filename: testFiles[0]),
|
||||
threshold: threshold)) {
|
||||
print("{$testFiles[0]} should match fangjun!");
|
||||
return;
|
||||
}
|
||||
|
||||
if (!manager.remove("fangjun")) {
|
||||
print("Failed to remove fangjun");
|
||||
return;
|
||||
}
|
||||
|
||||
if (manager.verify(
|
||||
name: "fangjun",
|
||||
embedding: computeEmbedding(extractor: extractor, filename: testFiles[0]),
|
||||
threshold: threshold)) {
|
||||
print("${testFiles[0]} should match no one!");
|
||||
return;
|
||||
}
|
||||
|
||||
if (manager.numSpeakers != 1) {
|
||||
print("There should only 1 speaker left.");
|
||||
return;
|
||||
}
|
||||
|
||||
extractor.free();
|
||||
manager.free();
|
||||
}
|
||||
17
dart-api-examples/speaker-identification/pubspec.yaml
Normal file
17
dart-api-examples/speaker-identification/pubspec.yaml
Normal file
@@ -0,0 +1,17 @@
|
||||
name: speaker_identification
|
||||
|
||||
description: >
|
||||
This example demonstrates how to use the Dart API for speaker identification.
|
||||
|
||||
version: 1.0.0
|
||||
|
||||
environment:
|
||||
sdk: ^3.4.0
|
||||
|
||||
dependencies:
|
||||
sherpa_onnx: ^1.10.20
|
||||
path: ^1.9.0
|
||||
args: ^2.5.0
|
||||
|
||||
dev_dependencies:
|
||||
lints: ^3.0.0
|
||||
19
dart-api-examples/speaker-identification/run-3d-speaker.sh
Executable file
19
dart-api-examples/speaker-identification/run-3d-speaker.sh
Executable file
@@ -0,0 +1,19 @@
|
||||
#!/usr/bin/env bash
|
||||
|
||||
set -ex
|
||||
|
||||
dart pub get
|
||||
|
||||
if [ ! -f ./3dspeaker_speech_eres2net_base_sv_zh-cn_3dspeaker_16k.onnx ]; then
|
||||
curl -SL -O https://github.com/k2-fsa/sherpa-onnx/releases/download/speaker-recongition-models/3dspeaker_speech_eres2net_base_sv_zh-cn_3dspeaker_16k.onnx
|
||||
fi
|
||||
|
||||
if [ ! -f ./sr-data/enroll/leijun-sr-1.wav ]; then
|
||||
curl -SL -o sr-data.tar.gz https://github.com/csukuangfj/sr-data/archive/refs/tags/v1.0.0.tar.gz
|
||||
tar xvf sr-data.tar.gz
|
||||
mv sr-data-1.0.0 sr-data
|
||||
fi
|
||||
|
||||
dart run \
|
||||
./bin/speaker_id.dart \
|
||||
--model ./3dspeaker_speech_eres2net_base_sv_zh-cn_3dspeaker_16k.onnx
|
||||
@@ -15,3 +15,7 @@
|
||||
|Non-Streaming speech recognition| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/non-streaming-asr)| macOS, Windows, Linux|
|
||||
|Text to speech| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/tts)| macOS, Windows, Linux|
|
||||
|Voice activity detection (VAD)| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/vad)| macOS, Windows, Linux|
|
||||
|Voice activity detection (VAD) with non-streaming speech recognition| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/vad-with-non-streaming-asr)| macOS, Windows, Linux|
|
||||
|Speaker identification and verification| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/speaker-identification)| macOS, Windows, Linux|
|
||||
|Audio tagging| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/audio-tagging)| macOS, Windows, Linux|
|
||||
|Keyword spotter| [Address](https://github.com/k2-fsa/sherpa-onnx/tree/master/dart-api-examples/keyword-spotter)| macOS, Windows, Linux|
|
||||
|
||||
@@ -107,7 +107,7 @@ public class SpeakerIdentification {
|
||||
|
||||
// test verify
|
||||
if (!manager.verify("fangjun", computeEmbedding(extractor, testFiles[0]), threshold)) {
|
||||
System.out.printf("testFiles[0] should match fangjun!");
|
||||
System.out.printf("%s should match fangjun!\n", testFiles[0]);
|
||||
return;
|
||||
}
|
||||
|
||||
|
||||
18
scripts/dart/speaker-id-pubspec.yaml
Normal file
18
scripts/dart/speaker-id-pubspec.yaml
Normal file
@@ -0,0 +1,18 @@
|
||||
name: speaker_identification
|
||||
|
||||
description: >
|
||||
This example demonstrates how to use the Dart API for speaker identification.
|
||||
|
||||
version: 1.0.0
|
||||
|
||||
environment:
|
||||
sdk: ^3.4.0
|
||||
|
||||
dependencies:
|
||||
sherpa_onnx:
|
||||
path: ../../flutter/sherpa_onnx
|
||||
path: ^1.9.0
|
||||
args: ^2.5.0
|
||||
|
||||
dev_dependencies:
|
||||
lints: ^3.0.0
|
||||
Reference in New Issue
Block a user