First of all, thank you for sharing your awesome works such as FastSpeech2, VocGAN, AdaSpeech and etcs.
It helped me a lot.
I leave this question due to my curiosity about duration predictor improvements.
A few months ago, evaluation performance of duration predictor seemed to be not good due to overfitting. (train error: below 0.01, but eval error: 0.5~0.6.)
But, now it has been drastically improved (eval error 0.5~0.6 -> 0.06 ~ 0.08).
If you don't mind, could you tell me what was the problem of your previous version of duration predictor?
Hello @rishikksh20 ,
First of all, thank you for sharing your awesome works such as FastSpeech2, VocGAN, AdaSpeech and etcs. It helped me a lot.
I leave this question due to my curiosity about duration predictor improvements. A few months ago, evaluation performance of duration predictor seemed to be not good due to overfitting. (train error: below 0.01, but eval error: 0.5~0.6.)
But, now it has been drastically improved (eval error 0.5~0.6 -> 0.06 ~ 0.08).
If you don't mind, could you tell me what was the problem of your previous version of duration predictor?
Always appreciate,