// sherpa-onnx/csrc/offline-recognizer-sense-voice-impl.h // // Copyright (c) 2022-2023 Xiaomi Corporation #ifndef SHERPA_ONNX_CSRC_OFFLINE_RECOGNIZER_SENSE_VOICE_IMPL_H_ #define SHERPA_ONNX_CSRC_OFFLINE_RECOGNIZER_SENSE_VOICE_IMPL_H_ #include #include #include #include #include #include "sherpa-onnx/csrc/offline-ctc-greedy-search-decoder.h" #include "sherpa-onnx/csrc/offline-model-config.h" #include "sherpa-onnx/csrc/offline-recognizer-impl.h" #include "sherpa-onnx/csrc/offline-recognizer.h" #include "sherpa-onnx/csrc/offline-sense-voice-model.h" #include "sherpa-onnx/csrc/pad-sequence.h" #include "sherpa-onnx/csrc/symbol-table.h" namespace sherpa_onnx { static OfflineRecognitionResult ConvertSenseVoiceResult( const OfflineCtcDecoderResult &src, const SymbolTable &sym_table, int32_t frame_shift_ms, int32_t subsampling_factor) { OfflineRecognitionResult r; r.tokens.reserve(src.tokens.size()); r.timestamps.reserve(src.timestamps.size()); std::string text; for (int32_t i = 4; i < src.tokens.size(); ++i) { auto sym = sym_table[src.tokens[i]]; text.append(sym); r.tokens.push_back(std::move(sym)); } r.text = std::move(text); float frame_shift_s = frame_shift_ms / 1000. * subsampling_factor; for (int32_t i = 4; i < src.timestamps.size(); ++i) { float time = frame_shift_s * (src.timestamps[i] - 4); r.timestamps.push_back(time); } r.words = std::move(src.words); // parse lang, emotion and event from tokens. if (src.tokens.size() >= 3) { r.lang = sym_table[src.tokens[0]]; r.emotion = sym_table[src.tokens[1]]; r.event = sym_table[src.tokens[2]]; } return r; } class OfflineRecognizerSenseVoiceImpl : public OfflineRecognizerImpl { public: explicit OfflineRecognizerSenseVoiceImpl( const OfflineRecognizerConfig &config) : OfflineRecognizerImpl(config), config_(config), symbol_table_(config_.model_config.tokens), model_(std::make_unique(config.model_config)) { const auto &meta_data = model_->GetModelMetadata(); if (config.decoding_method == "greedy_search") { decoder_ = std::make_unique(meta_data.blank_id); } else { SHERPA_ONNX_LOGE("Only greedy_search is supported at present. Given %s", config.decoding_method.c_str()); exit(-1); } InitFeatConfig(); } template OfflineRecognizerSenseVoiceImpl(Manager *mgr, const OfflineRecognizerConfig &config) : OfflineRecognizerImpl(mgr, config), config_(config), symbol_table_(mgr, config_.model_config.tokens), model_(std::make_unique(mgr, config.model_config)) { const auto &meta_data = model_->GetModelMetadata(); if (config.decoding_method == "greedy_search") { decoder_ = std::make_unique(meta_data.blank_id); } else { SHERPA_ONNX_LOGE("Only greedy_search is supported at present. Given %s", config.decoding_method.c_str()); exit(-1); } InitFeatConfig(); } std::unique_ptr CreateStream() const override { return std::make_unique(config_.feat_config); } void DecodeStreams(OfflineStream **ss, int32_t n) const override { if (n == 1) { DecodeOneStream(ss[0]); return; } const auto &meta_data = model_->GetModelMetadata(); // 1. Apply LFR // 2. Apply CMVN // // Please refer to // https://static.googleusercontent.com/media/research.google.com/en//pubs/archive/45555.pdf // for what LFR means // // "Lower Frame Rate Neural Network Acoustic Models" auto memory_info = Ort::MemoryInfo::CreateCpu(OrtDeviceAllocator, OrtMemTypeDefault); std::vector features; features.reserve(n); int32_t feat_dim = config_.feat_config.feature_dim * meta_data.window_size; std::vector> features_vec(n); std::vector features_length_vec(n); for (int32_t i = 0; i != n; ++i) { std::vector f = ss[i]->GetFrames(); f = ApplyLFR(f); ApplyCMVN(&f); int32_t num_frames = f.size() / feat_dim; features_vec[i] = std::move(f); features_length_vec[i] = num_frames; std::array shape = {num_frames, feat_dim}; Ort::Value x = Ort::Value::CreateTensor( memory_info, features_vec[i].data(), features_vec[i].size(), shape.data(), shape.size()); features.push_back(std::move(x)); } std::vector features_pointer(n); for (int32_t i = 0; i != n; ++i) { features_pointer[i] = &features[i]; } std::array features_length_shape = {n}; Ort::Value x_length = Ort::Value::CreateTensor( memory_info, features_length_vec.data(), n, features_length_shape.data(), features_length_shape.size()); // Caution(fangjun): We cannot pad it with log(eps), // i.e., -23.025850929940457f Ort::Value x = PadSequence(model_->Allocator(), features_pointer, 0); int32_t language = 0; if (config_.model_config.sense_voice.language.empty()) { language = 0; } else if (meta_data.lang2id.count( config_.model_config.sense_voice.language)) { language = meta_data.lang2id.at(config_.model_config.sense_voice.language); } else { SHERPA_ONNX_LOGE("Unknown language: %s. Use 0 instead.", config_.model_config.sense_voice.language.c_str()); } std::vector language_array(n); std::fill(language_array.begin(), language_array.end(), language); std::vector text_norm_array(n); std::fill(text_norm_array.begin(), text_norm_array.end(), config_.model_config.sense_voice.use_itn ? meta_data.with_itn_id : meta_data.without_itn_id); Ort::Value language_tensor = Ort::Value::CreateTensor( memory_info, language_array.data(), n, features_length_shape.data(), features_length_shape.size()); Ort::Value text_norm_tensor = Ort::Value::CreateTensor( memory_info, text_norm_array.data(), n, features_length_shape.data(), features_length_shape.size()); Ort::Value logits{nullptr}; try { logits = model_->Forward(std::move(x), std::move(x_length), std::move(language_tensor), std::move(text_norm_tensor)); } catch (const Ort::Exception &ex) { SHERPA_ONNX_LOGE("\n\nCaught exception:\n\n%s\n\nReturn an empty result", ex.what()); return; } // decoder_->Decode() requires that logits_length is of dtype int64 std::vector features_length_vec_64; features_length_vec_64.reserve(n); for (auto i : features_length_vec) { i += 4; features_length_vec_64.push_back(i); } Ort::Value logits_length = Ort::Value::CreateTensor( memory_info, features_length_vec_64.data(), n, features_length_shape.data(), features_length_shape.size()); auto results = decoder_->Decode(std::move(logits), std::move(logits_length)); int32_t frame_shift_ms = 10; int32_t subsampling_factor = meta_data.window_shift; for (int32_t i = 0; i != n; ++i) { auto r = ConvertSenseVoiceResult(results[i], symbol_table_, frame_shift_ms, subsampling_factor); r.text = ApplyInverseTextNormalization(std::move(r.text)); ss[i]->SetResult(r); } } OfflineRecognizerConfig GetConfig() const override { return config_; } private: void DecodeOneStream(OfflineStream *s) const { const auto &meta_data = model_->GetModelMetadata(); auto memory_info = Ort::MemoryInfo::CreateCpu(OrtDeviceAllocator, OrtMemTypeDefault); int32_t feat_dim = config_.feat_config.feature_dim * meta_data.window_size; std::vector f = s->GetFrames(); f = ApplyLFR(f); ApplyCMVN(&f); int32_t num_frames = f.size() / feat_dim; std::array shape = {1, num_frames, feat_dim}; Ort::Value x = Ort::Value::CreateTensor(memory_info, f.data(), f.size(), shape.data(), shape.size()); int64_t scale_shape = 1; Ort::Value x_length = Ort::Value::CreateTensor(memory_info, &num_frames, 1, &scale_shape, 1); int32_t language = 0; if (config_.model_config.sense_voice.language.empty()) { language = 0; } else if (meta_data.lang2id.count( config_.model_config.sense_voice.language)) { language = meta_data.lang2id.at(config_.model_config.sense_voice.language); } else { SHERPA_ONNX_LOGE("Unknown language: %s. Use 0 instead.", config_.model_config.sense_voice.language.c_str()); } int32_t text_norm = config_.model_config.sense_voice.use_itn ? meta_data.with_itn_id : meta_data.without_itn_id; Ort::Value language_tensor = Ort::Value::CreateTensor(memory_info, &language, 1, &scale_shape, 1); Ort::Value text_norm_tensor = Ort::Value::CreateTensor(memory_info, &text_norm, 1, &scale_shape, 1); Ort::Value logits{nullptr}; try { logits = model_->Forward(std::move(x), std::move(x_length), std::move(language_tensor), std::move(text_norm_tensor)); } catch (const Ort::Exception &ex) { SHERPA_ONNX_LOGE("\n\nCaught exception:\n\n%s\n\nReturn an empty result", ex.what()); return; } int64_t new_num_frames = num_frames + 4; Ort::Value logits_length = Ort::Value::CreateTensor( memory_info, &new_num_frames, 1, &scale_shape, 1); auto results = decoder_->Decode(std::move(logits), std::move(logits_length)); int32_t frame_shift_ms = 10; int32_t subsampling_factor = meta_data.window_shift; auto r = ConvertSenseVoiceResult(results[0], symbol_table_, frame_shift_ms, subsampling_factor); r.text = ApplyInverseTextNormalization(std::move(r.text)); s->SetResult(r); } void InitFeatConfig() { const auto &meta_data = model_->GetModelMetadata(); config_.feat_config.normalize_samples = meta_data.normalize_samples; config_.feat_config.window_type = "hamming"; config_.feat_config.high_freq = 0; config_.feat_config.snip_edges = true; } std::vector ApplyLFR(const std::vector &in) const { const auto &meta_data = model_->GetModelMetadata(); int32_t lfr_window_size = meta_data.window_size; int32_t lfr_window_shift = meta_data.window_shift; int32_t in_feat_dim = config_.feat_config.feature_dim; int32_t in_num_frames = in.size() / in_feat_dim; int32_t out_num_frames = (in_num_frames - lfr_window_size) / lfr_window_shift + 1; int32_t out_feat_dim = in_feat_dim * lfr_window_size; std::vector out(out_num_frames * out_feat_dim); const float *p_in = in.data(); float *p_out = out.data(); for (int32_t i = 0; i != out_num_frames; ++i) { std::copy(p_in, p_in + out_feat_dim, p_out); p_out += out_feat_dim; p_in += lfr_window_shift * in_feat_dim; } return out; } void ApplyCMVN(std::vector *v) const { const auto &meta_data = model_->GetModelMetadata(); const std::vector &neg_mean = meta_data.neg_mean; const std::vector &inv_stddev = meta_data.inv_stddev; int32_t dim = neg_mean.size(); int32_t num_frames = v->size() / dim; float *p = v->data(); for (int32_t i = 0; i != num_frames; ++i) { for (int32_t k = 0; k != dim; ++k) { p[k] = (p[k] + neg_mean[k]) * inv_stddev[k]; } p += dim; } } OfflineRecognizerConfig config_; SymbolTable symbol_table_; std::unique_ptr model_; std::unique_ptr decoder_; }; } // namespace sherpa_onnx #endif // SHERPA_ONNX_CSRC_OFFLINE_RECOGNIZER_SENSE_VOICE_IMPL_H_