Closed bopjesvla closed 4 years ago
Can you reproduce this on 0.5.0?
Tried this myself on a stock 0.5.1 with only librivox CLEAN sets only. Can't reproduce
@rhamnett are you able spot decrease in WER when you are training in 0.5.0. I am also facing an increase in WER when training on a custom dataset in 0.4.1
Might be wrong but seems the 0.5.1 was trained with some extra data set missing and will be corrected in 0.6.0
@bopjesvla Is that still an issue with 0.6 ?
I won't be able to check this in the near future. Since it couldn't be reproduced by @rhamnett in 0.5.1, I'll close the issue.
This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs.
This bug was discussed here with @lissyx: https://discourse.mozilla.org/t/wer-shoots-up-when-retraining-the-pretrained-model-for-an-additional-epoch-on-libri/41062
The relevant output:
If epoch is set to 0, the WER is 0.08, which is about expected. Why would the WER shoot up to 0.7 when training for one more epoch?
I understand that the pretrained model was trained on more than just LibriSpeech, but the difference is incredibly large. The reason I’m asking is that I’m seeing similar increases in WER when I continue to train the model on another dataset. The WER of the non-finetuned pretrained model on this dataset is 0.11, but when I train the model on the dataset, the WER immediately jumps up to 0.4 after one epoch.
On the forums, Lissyx posted the following reply:
This suggestion was followed, but to no avail. Setting the learning rate to 0.00001 results in a WER between 0.97 and 0.99 after one epoch. The same is true for a value of 0.000001. To verify nothing else changed, I reran the script with the original learning rate, 0.0001, resulting in a WER of 0.61.
The relevant output for lr = 0.00001:
0.000001:
Training for two epochs with the original, higher learning rate also results in a WER of 0.98.