This repository has been archived on 2025-08-26. You can view files and clone it, but cannot push or open issues or pull requests.
Files
enginex_bi_series-sherpa-onnx/pascal-api-examples/vad-with-non-streaming-asr
Fangjun Kuang 3bf986d08d Support non-streaming zipformer CTC ASR models (#2340)
This PR adds support for non-streaming Zipformer CTC ASR models across 
multiple language bindings, WebAssembly, examples, and CI workflows.

- Introduces a new OfflineZipformerCtcModelConfig in C/C++, Python, Swift, Java, Kotlin, Go, Dart, Pascal, and C# APIs
- Updates initialization, freeing, and recognition logic to include Zipformer CTC in WASM and Node.js
- Adds example scripts and CI steps for downloading, building, and running Zipformer CTC models

Model doc is available at
https://k2-fsa.github.io/sherpa/onnx/pretrained_models/offline-ctc/icefall/zipformer.html
2025-07-04 15:57:07 +08:00
..

Introduction

This directory contains examples for how to use the VAD (voice activity detection) with non-streaming speech recognition models.

Directory Description
run-vad-with-dolphin-ctc.sh It shows how to use the VAD + Dolphin for speech recognition.
run-vad-with-whisper.sh It shows how to use the VAD + Whisper for speech recognition.
run-vad-with-sense-voice.sh It shows how to use the VAD + SenseVoice for speech recognition.
run-vad-with-moonshine.sh It shows how to use the VAD + Moonshine for speech recognition.

Please refer to non-streaming-asr for more kinds of non-streaming models.