Open bmwshop opened 5 years ago
What was the output of greedy decoder?
I have the same problem.
I guess the problem is ctc_decoder.When i made the language model from native client provided by original deepspeech implementation, the results were impressive but the inference speed was very low. But when i built the language model with OpenSeq2Seq toolkit the inference speed is high but wer is also high. I am getting the results like above. e.g on forms that selemtwetysevenfeightenhoe
On certain samples, intermittently (but more likely on longer ones), when we inference with KenLM, we end up with gibberish at the end of the end of the transcript, eg.
"four one seven crivenuehioforfivzfofofivtresevonfozodichargdignowetnpscosinollservicetdawitworefis...."
this contributes to an overall lower WER score with an LM.