232 lines
6.3 KiB
Python
Executable File
232 lines
6.3 KiB
Python
Executable File
#!/usr/bin/env python3
|
|
|
|
# Real-time speech recognition from a microphone with sherpa-onnx Python API
|
|
#
|
|
# Please refer to
|
|
# https://k2-fsa.github.io/sherpa/onnx/pretrained_models/index.html
|
|
# to download pre-trained models
|
|
|
|
import argparse
|
|
import sys
|
|
from pathlib import Path
|
|
|
|
from typing import List
|
|
import sentencepiece as spm
|
|
|
|
try:
|
|
import sounddevice as sd
|
|
except ImportError:
|
|
print("Please install sounddevice first. You can use")
|
|
print()
|
|
print(" pip install sounddevice")
|
|
print()
|
|
print("to install it")
|
|
sys.exit(-1)
|
|
|
|
import sherpa_onnx
|
|
|
|
|
|
def assert_file_exists(filename: str):
|
|
assert Path(filename).is_file(), (
|
|
f"{filename} does not exist!\n"
|
|
"Please refer to "
|
|
"https://k2-fsa.github.io/sherpa/onnx/pretrained_models/index.html to download it"
|
|
)
|
|
|
|
|
|
def get_args():
|
|
parser = argparse.ArgumentParser(
|
|
formatter_class=argparse.ArgumentDefaultsHelpFormatter
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--tokens",
|
|
type=str,
|
|
required=True,
|
|
help="Path to tokens.txt",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--encoder",
|
|
type=str,
|
|
required=True,
|
|
help="Path to the encoder model",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--decoder",
|
|
type=str,
|
|
required=True,
|
|
help="Path to the decoder model",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--joiner",
|
|
type=str,
|
|
help="Path to the joiner model",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--decoding-method",
|
|
type=str,
|
|
default="greedy_search",
|
|
help="Valid values are greedy_search and modified_beam_search",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--max-active-paths",
|
|
type=int,
|
|
default=4,
|
|
help="""Used only when --decoding-method is modified_beam_search.
|
|
It specifies number of active paths to keep during decoding.
|
|
""",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--provider",
|
|
type=str,
|
|
default="cpu",
|
|
help="Valid values: cpu, cuda, coreml",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--bpe-model",
|
|
type=str,
|
|
default="",
|
|
help="""
|
|
Path to bpe.model, it will be used to tokenize contexts biasing phrases.
|
|
Used only when --decoding-method=modified_beam_search
|
|
""",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--modeling-unit",
|
|
type=str,
|
|
default="char",
|
|
help="""
|
|
The type of modeling unit, it will be used to tokenize contexts biasing phrases.
|
|
Valid values are bpe, bpe+char, char.
|
|
Note: the char here means characters in CJK languages.
|
|
Used only when --decoding-method=modified_beam_search
|
|
""",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--contexts",
|
|
type=str,
|
|
default="",
|
|
help="""
|
|
The context list, it is a string containing some words/phrases separated
|
|
with /, for example, 'HELLO WORLD/I LOVE YOU/GO AWAY".
|
|
Used only when --decoding-method=modified_beam_search
|
|
""",
|
|
)
|
|
|
|
parser.add_argument(
|
|
"--context-score",
|
|
type=float,
|
|
default=1.5,
|
|
help="""
|
|
The context score of each token for biasing word/phrase. Used only if
|
|
--contexts is given.
|
|
Used only when --decoding-method=modified_beam_search
|
|
""",
|
|
)
|
|
|
|
return parser.parse_args()
|
|
|
|
|
|
def create_recognizer(args):
|
|
assert_file_exists(args.encoder)
|
|
assert_file_exists(args.decoder)
|
|
assert_file_exists(args.joiner)
|
|
assert_file_exists(args.tokens)
|
|
# Please replace the model files if needed.
|
|
# See https://k2-fsa.github.io/sherpa/onnx/pretrained_models/index.html
|
|
# for download links.
|
|
recognizer = sherpa_onnx.OnlineRecognizer.from_transducer(
|
|
tokens=args.tokens,
|
|
encoder=args.encoder,
|
|
decoder=args.decoder,
|
|
joiner=args.joiner,
|
|
num_threads=1,
|
|
sample_rate=16000,
|
|
feature_dim=80,
|
|
decoding_method=args.decoding_method,
|
|
max_active_paths=args.max_active_paths,
|
|
provider=args.provider,
|
|
context_score=args.context_score,
|
|
)
|
|
return recognizer
|
|
|
|
|
|
def encode_contexts(args, contexts: List[str]) -> List[List[int]]:
|
|
sp = None
|
|
if "bpe" in args.modeling_unit:
|
|
assert_file_exists(args.bpe_model)
|
|
sp = spm.SentencePieceProcessor()
|
|
sp.load(args.bpe_model)
|
|
tokens = {}
|
|
with open(args.tokens, "r", encoding="utf-8") as f:
|
|
for line in f:
|
|
toks = line.strip().split()
|
|
assert len(toks) == 2, len(toks)
|
|
assert toks[0] not in tokens, f"Duplicate token: {toks} "
|
|
tokens[toks[0]] = int(toks[1])
|
|
return sherpa_onnx.encode_contexts(
|
|
modeling_unit=args.modeling_unit,
|
|
contexts=contexts,
|
|
sp=sp,
|
|
tokens_table=tokens,
|
|
)
|
|
|
|
|
|
def main():
|
|
args = get_args()
|
|
|
|
devices = sd.query_devices()
|
|
if len(devices) == 0:
|
|
print("No microphone devices found")
|
|
sys.exit(0)
|
|
|
|
print(devices)
|
|
default_input_device_idx = sd.default.device[0]
|
|
print(f'Use default device: {devices[default_input_device_idx]["name"]}')
|
|
|
|
contexts_list = []
|
|
contexts = [x.strip().upper() for x in args.contexts.split("/") if x.strip()]
|
|
if contexts:
|
|
print(f"Contexts list: {contexts}")
|
|
contexts_list = encode_contexts(args, contexts)
|
|
|
|
recognizer = create_recognizer(args)
|
|
print("Started! Please speak")
|
|
|
|
# The model is using 16 kHz, we use 48 kHz here to demonstrate that
|
|
# sherpa-onnx will do resampling inside.
|
|
sample_rate = 48000
|
|
samples_per_read = int(0.1 * sample_rate) # 0.1 second = 100 ms
|
|
last_result = ""
|
|
if contexts_list:
|
|
stream = recognizer.create_stream(contexts_list=contexts_list)
|
|
else:
|
|
stream = recognizer.create_stream()
|
|
with sd.InputStream(channels=1, dtype="float32", samplerate=sample_rate) as s:
|
|
while True:
|
|
samples, _ = s.read(samples_per_read) # a blocking read
|
|
samples = samples.reshape(-1)
|
|
stream.accept_waveform(sample_rate, samples)
|
|
while recognizer.is_ready(stream):
|
|
recognizer.decode_stream(stream)
|
|
result = recognizer.get_result(stream)
|
|
if last_result != result:
|
|
last_result = result
|
|
print("\r{}".format(result), end="", flush=True)
|
|
|
|
|
|
if __name__ == "__main__":
|
|
try:
|
|
main()
|
|
except KeyboardInterrupt:
|
|
print("\nCaught Ctrl + C. Exiting")
|