Closed AigizK closed 15 hours ago
All training details is mentioned in our paper.
And you could simply train your own model for a new language:
For Base model (multilingual, ~300M), we use <50K hours for each language (EN ZH) For Small model (e.g. Chinese-only, ~150M), we have made it work with just 1K hours data, config. mentioned in our paper also
Just one thing, the training would take a long time, especially for E2 TTS (if you choose) And be patient, 8 x RTX3090 small model for one week (200~400K updates to hear something reasonable) 8 x A100 for base model similarly.
Thank you for your work. I would like to inquire about the possibility of training for a new language. If this is feasible, could you please provide more details on the following:
Your insights on this matter would be greatly appreciated. Thank you in advance for your assistance.