Fix running MeloTTS models on GPU. (#1379)
We need to use opset 18 to export the model to onnx.
This commit is contained in:
@@ -229,7 +229,7 @@ def main():
|
||||
|
||||
torch_model = ModelWrapper(model)
|
||||
|
||||
opset_version = 13
|
||||
opset_version = 18
|
||||
x = torch.randint(low=0, high=10, size=(60,), dtype=torch.int64)
|
||||
print(x.shape)
|
||||
x_lengths = torch.tensor([x.size(0)], dtype=torch.int64)
|
||||
|
||||
Reference in New Issue
Block a user