Closed edosyhptra closed 3 years ago
- if valid_wer for ex "8.314", is that means 8% of wer or 80%?
8% WER not 80%. Try using a LM should reduce the WER a bit.
- Given the valid_wer and valid_raw_wer ( "valid_wer": "8.314", "valid_raw_wer": "8.314) , it occurs the same in every validation process. Is it because I am not using lm decoding on args?
Yes. raw_wer is the CTC wer before LM decoding. if you do not use an LM, wer=raw_wer
- if valid_wer for ex "8.314", is that means 8% of wer or 80%?
8% WER not 80%. Try using a LM should reduce the WER a bit.
I see. I will try it!
Thanks for the answer! really appreciate
❓ Questions and Help
I have been pretrained wav2vec2 base model with my own dataset in Bahasa Indonesia for approximately 300 hours and got 56% valid accuracy. Now still running finetuning with around 50h dataset from the pretrained model and got valid_wer: 8.736 on epoch 37.
What is your question?
Im not sure whether my pretrained model is overfit. I have been searched for explanation for loss0, loss1, code_perplexity, etc to understand how good my pretrained model learned.
Given the pretrain and finetune log,
Pretrain log
Finetune log