Open ArtemisZGL opened 4 years ago
Multi-speaker is supported as well. For instance, you might collect 8 speakers and one hour for each corpus and record the directory in scripts/train_tacotron2.sh
. Then the total amounts of data might help.
@begeekmyfriend Thanks for your reply, but one speaker in libritts just 10-20 min, if i just use about 10 speaker of this, will it work?
I have no idea about your circumstances but it seems too little for everyone's corpus. You might try yourself but the evaluation can not be warranted.
Thankyou so much sir, for such an amazing work. @begeekmyfriend
Thanks for your job first, I want to know that how many data should i use to train in your repo. I want to use the cmu arctic dataset for training english tts, but there is about one hour for a speaker, can it work to train in you repo?Because i use the nvidia repo to train but the result is bad. Also, the result is very different for different batch size. I also use some of the libritts data to train on nvdia repo, about 3hour for 10 speaker, but the result is quiet bad too. Do you have some idea about how to train on small dataset ?