Closed NAM-hj closed 4 years ago
how optimization proceeds depends on lots of factors: 1) your model architecture 2) your dataset 3) the task you are trying to solve 4) your batch size + lr (i.e. how many gpus you are using)
looking at the graph of your loss, it seems to be working well. its possible that the model has already learned good representations, which you can test by trying to use them for e.g. timit or zerospeech or something like this. otherwise you can try to train with a higher learning rate, use more gpus to train, train for longer, etc
❓ Questions and Help
What is your question?
[How can I train well a new model with CLI tools?]
I tried to train a wav2vec model for the Librispeech. But I couldn't get the well-trained model, because the loss was started at 5.xx but still 2.2x on epoch 47. What should I try to get lower loss?
What have you tried?
python examples/wav2vec/wav2vec_manifest.py ./data/librispeech --dest ./data/prepared_full --ext flac
→ I use whole files of the Librispeech.→ The loss log of the above code is presented below.
(+) Some changes like '256 to 512' and 'large version of wav2vec which described in the paper' are insignificant.
What's your environment?
pip
, source): install with sourceapex and pyarrow are installed too.