Use espeak-ng for coqui-ai/TTS VITS English models. (#466)

This commit is contained in:
Fangjun Kuang
2023-12-06 11:00:38 +08:00
committed by GitHub
parent 3b90e85ef2
commit 23cf92daf7
10 changed files with 230 additions and 93 deletions

View File

@@ -0,0 +1,34 @@
// sherpa-onnx/csrc/offline-tts-vits-model-metadata.h
//
// Copyright (c) 2023 Xiaomi Corporation
#ifndef SHERPA_ONNX_CSRC_OFFLINE_TTS_VITS_MODEL_METADATA_H_
#define SHERPA_ONNX_CSRC_OFFLINE_TTS_VITS_MODEL_METADATA_H_
#include <cstdint>
#include <string>
namespace sherpa_onnx {
struct OfflineTtsVitsModelMetaData {
int32_t sample_rate;
int32_t add_blank = 0;
int32_t num_speakers = 0;
std::string punctuations;
std::string language;
std::string voice;
bool is_piper = false;
bool is_coqui = false;
// the following options are for models from coqui-ai/TTS
int32_t blank_id = 0;
int32_t bos_id = 0;
int32_t eos_id = 0;
int32_t use_eos_bos = 0;
};
} // namespace sherpa_onnx
#endif // SHERPA_ONNX_CSRC_OFFLINE_TTS_VITS_MODEL_METADATA_H_