Open chynphh opened 4 years ago
Is the loss still going down and model is not overfitting, i.e. generalization error is not increasing?
I also use multi-speaker data, include half of biaobei, a private male datasets(5000 sentences), and other 4 private small datasets, totally 6 speakers. When I use this datasets to train mellotron, I can't get a good alignment, the alignment like the below figure But while I only use the single speaker dataset bioabei, the alignment is good. So I want to know, what's your mandarin multi-speaker data look like, how many speakers? And 6 speakers is enough? @chynphh
When using multi-speaker data, can not distinguish between male and female, and there are only slight differences between different speakers. Current training step is 32K. Is this normal? The language is mandarin.