Open liuxubo717 opened 3 years ago
Thanks for great work. I have the same question with @liuxubo717 , I compare with the tacotron2, find that its postnet is "torch.nn.Conv1d", I don't understand that using the "Bi-GRU" in your postnet, does it better than the postnet of Tacotron2? can you explain than? Thanks.
Many thanks for your great work.
I have a question about the Post-Net after the Mel Linear layer, I find in the inference stage, you use the mel_pred as output rather than the postnet_pred, which is shown as follows (in synthesis,py):
Also, when I run the code, I find that the post_mel_loss is always bigger than the mel_loss, which means the post_Net module doesn't work as expected, right? I think it is conflicted with the Post-Net module used in Tacotron and the TTS-Transformer original paper. I am a bit confused, can you explain it to me? Many thanks!