Plachtaa / VITS-fast-fine-tuning

This repo is a pipeline of VITS finetuning for fast speaker adaptation TTS, and many-to-many voice conversion
Apache License 2.0
4.69k stars 703 forks source link

drop_speaker_embed在第一次启动时是True, 继续训练时改成了False,请问drop_speaker_embed是什么意思, 感谢~ #515

Open mikeyang01 opened 9 months ago

mikeyang01 commented 9 months ago

英文文档里, 第一次执行: Run python finetune_speaker_v2.py -m ./OUTPUT_MODEL --max_epochs "{Maximum_epochs}" --drop_speaker_embed True Do replace {Maximum_epochs} with your desired number of epochs. Empirically, 100 or more is recommended.

继续执行: To continue training on previous checkpoint, change the training command to: python finetune_speaker_v2.py -m ./OUTPUT_MODEL --max_epochs "{Maximum_epochs}" --drop_speaker_embed False --cont True. Before you do this, make sure you have previous G_latest.pth and D_latest.pth under ./OUTPUT_MODEL/ directory.