Add C++ runtime for silero_vad with RKNN (#2078)

This commit is contained in:
Fangjun Kuang
2025-04-01 15:56:56 +08:00
committed by GitHub
parent 0703bc1b86
commit 8e51a97550
12 changed files with 867 additions and 16 deletions

View File

@@ -5,15 +5,94 @@ import onnx
import torch
from onnxsim import simplify
import torch
from torch import Tensor
def simple_pad(x: Tensor, pad: int) -> Tensor:
# _0 = torch.slice(torch.slice(torch.slice(x), 1), 2, 1, torch.add(1, pad))
_0 = x[:, :, 1 : 1 + pad]
left_pad = torch.flip(_0, [-1])
# _1 = torch.slice(torch.slice(torch.slice(x), 1), 2, torch.sub(-1, pad), -1)
_1 = x[:, :, (-1 - pad) : -1]
right_pad = torch.flip(_1, [-1])
_2 = torch.cat([left_pad, x, right_pad], 2)
return _2
class MyModule(torch.nn.Module):
def __init__(self, m):
super().__init__()
self.m = m
def adaptive_normalization_forward(self, spect):
m = self.m._model.adaptive_normalization
_0 = simple_pad
# Note(fangjun): rknn uses fp16 by default, whose max value is 65504
# so we need to re-write the computation for spect0
# spect0 = torch.log1p(torch.mul(spect, 1048576))
spect0 = torch.log1p(spect) + 13.86294
_1 = torch.eq(len(spect0.shape), 2)
if _1:
_2 = torch.unsqueeze(spect0, 0)
spect1 = _2
else:
spect1 = spect0
mean = torch.mean(spect1, [1], True)
to_pad = m.to_pad
mean0 = _0(
mean,
to_pad,
)
filter_ = m.filter_
mean1 = torch.conv1d(mean0, filter_)
mean_mean = torch.mean(mean1, [-1], True)
spect2 = torch.add(spect1, torch.neg(mean_mean))
return spect2
def forward(self, x: torch.Tensor, h: torch.Tensor, c: torch.Tensor):
m = self.m._model
feature_extractor = m.feature_extractor
x0 = (feature_extractor).forward(
x,
)
norm = self.adaptive_normalization_forward(x0)
x1 = torch.cat([x0, norm], 1)
first_layer = m.first_layer
x2 = (first_layer).forward(
x1,
)
encoder = m.encoder
x3 = (encoder).forward(
x2,
)
decoder = m.decoder
x4, h0, c0, = (decoder).forward(
x3,
h,
c,
)
_0 = torch.mean(torch.squeeze(x4, 1), [1])
out = torch.unsqueeze(_0, 1)
return (out, h0, c0)
@torch.no_grad()
def main():
m = torch.jit.load("./silero_vad.jit")
m = MyModule(m)
x = torch.rand((1, 512), dtype=torch.float32)
h = torch.rand((2, 1, 64), dtype=torch.float32)
c = torch.rand((2, 1, 64), dtype=torch.float32)
m = torch.jit.script(m)
torch.onnx.export(
m._model,
m,
(x, h, c),
"m.onnx",
input_names=["x", "h", "c"],

View File

@@ -1,5 +1,5 @@
#!/usr/bin/env python3
# Copyright 2024 Xiaomi Corp. (authors: Fangjun Kuang)
# Copyright 2025 Xiaomi Corp. (authors: Fangjun Kuang)
import onnxruntime
import onnx

View File

@@ -0,0 +1,141 @@
#!/usr/bin/env python3
# Copyright 2025 Xiaomi Corp. (authors: Fangjun Kuang)
# Please run this file on your rk3588 board
try:
from rknnlite.api import RKNNLite
except:
print("Please run this file on your board (linux + aarch64 + npu)")
print("You need to install rknn_toolkit_lite2")
print(
" from https://github.com/airockchip/rknn-toolkit2/tree/master/rknn-toolkit-lite2/packages"
)
print(
"https://github.com/airockchip/rknn-toolkit2/blob/v2.1.0/rknn-toolkit-lite2/packages/rknn_toolkit_lite2-2.1.0-cp310-cp310-linux_aarch64.whl"
)
print("is known to work")
raise
import time
from pathlib import Path
from typing import Tuple
import numpy as np
import soundfile as sf
def load_audio(filename: str) -> Tuple[np.ndarray, int]:
data, sample_rate = sf.read(
filename,
always_2d=True,
dtype="float32",
)
data = data[:, 0] # use only the first channel
samples = np.ascontiguousarray(data)
return samples, sample_rate
def init_model(filename, target_platform="rk3588"):
if not Path(filename).is_file():
exit(f"{filename} does not exist")
rknn_lite = RKNNLite(verbose=False)
ret = rknn_lite.load_rknn(path=filename)
if ret != 0:
exit(f"Load model {filename} failed!")
ret = rknn_lite.init_runtime(core_mask=RKNNLite.NPU_CORE_0)
if ret != 0:
exit(f"Failed to init rknn runtime for {filename}")
return rknn_lite
class RKNNModel:
def __init__(self, model: str, target_platform="rk3588"):
self.model = init_model(model)
def release(self):
self.model.release()
def __call__(self, x: np.ndarray, h: np.ndarray, c: np.ndarray):
"""
Args:
x: (1, 512), np.float32
h: (2, 1, 64), np.float32
c: (2, 1, 64), np.float32
Returns:
prob:
next_h:
next_c
"""
out, next_h, next_c = self.model.inference(inputs=[x, h, c])
return out.item(), next_h, next_c
def main():
model = RKNNModel(model="./m.rknn")
for i in range(1):
test(model)
def test(model):
print("started")
start = time.time()
samples, sample_rate = load_audio("./lei-jun-test.wav")
assert sample_rate == 16000, sample_rate
window_size = 512
h = np.zeros((2, 1, 64), dtype=np.float32)
c = np.zeros((2, 1, 64), dtype=np.float32)
threshold = 0.5
num_windows = samples.shape[0] // window_size
out = []
for i in range(num_windows):
print(i, num_windows)
this_samples = samples[i * window_size : (i + 1) * window_size]
prob, h, c = model(this_samples[None], h, c)
out.append(prob > threshold)
min_speech_duration = 0.25 * sample_rate / window_size
min_silence_duration = 0.25 * sample_rate / window_size
result = []
last = -1
for k, f in enumerate(out):
if f >= threshold:
if last == -1:
last = k
elif last != -1:
if k - last > min_speech_duration:
result.append((last, k))
last = -1
if last != -1 and k - last > min_speech_duration:
result.append((last, k))
if not result:
print("Empty for ./lei-jun-test.wav")
return
print(result)
final = [result[0]]
for r in result[1:]:
f = final[-1]
if r[0] - f[1] < min_silence_duration:
final[-1] = (f[0], r[1])
else:
final.append(r)
for f in final:
start = f[0] * window_size / sample_rate
end = f[1] * window_size / sample_rate
print("{:.3f} -- {:.3f}".format(start, end))
if __name__ == "__main__":
main()

View File

@@ -97,10 +97,13 @@ def main():
h, c = model.get_init_states()
window_size = 512
num_windows = samples.shape[0] // window_size
for i in range(num_windows):
start = i * window_size
end = start + window_size
p, h, c = model(samples[start:end], h, c)
probs.append(p[0].item())
threshold = 0.5