OlaWod / FreeVC

FreeVC: Towards High-Quality Text-Free One-Shot Voice Conversion
MIT License
602 stars 111 forks source link

notebook Google Colab #43

Open HobisPL opened 1 year ago

HobisPL commented 1 year ago

Can you create a training model notebook on Google Colab? I tried to do it, but something is not working.

OlaWod commented 1 year ago

I'm afraid it is not practical to train on Colab, as Google Drive only has 15 GB free space.

# set the colab device to GPU
!git clone https://github.com/OlaWod/FreeVC.git
cd FreeVC
# do the data preprocessing according to your need
!python train.py -c configs/freevc.json -m freevc
HobisPL commented 1 year ago

Google Colab has around 60GB, Google Drive has 15GB. I mainly train other technologies from Nvidia (Tacotron 2, Flowtron, Rad-TTS) and Diff-SVC on Google Colab. I wanted to see how you would perform on the Polish language.

OlaWod commented 1 year ago

To be honest I feel a bit lazy to create a colab notebook. Isn't it feasible to just git clone this repo, and run the commands in the README with a "!" symbol in front of the command in colab? (I did not try but I thought it could work?) And btw the default SR augmentation will create 24x original data size. And if you only want to do inference I've offered a demo in huggingface.

steven850 commented 1 year ago

Yeah 60GB wouldn't cut it. VCTK with SR comes to 630GB after preprocess plus the Wavlm model etc... you need at least 700GB to train.

rbychn commented 1 year ago

And if you only want to do inference I've offered a demo in huggingface.

Could you please add a wiki or readme on how to get that UI on HuggingFace working locally? Would be great if you could also add UI for training in there so everything can be done locally, easily.

OlaWod commented 1 year ago

And if you only want to do inference I've offered a demo in huggingface.

Could you please add a wiki or readme on how to get that UI on HuggingFace working locally? Would be great if you could also add UI for training in there so everything can be done locally, easily.

for inference you can clone the huggingface repo and run app.py and go to http://localhost:7860/ in the browser. (gradio quickstart fyi) for training, sorry i'm lazy to do that.