Closed lomizandtyd closed 1 year ago
1 there are three losses in wenet-rnnt training style。 So you can also ues ctc + rescoring for decoding
2
3 fuison here means ctc probs fusion with transducer when decode , not with lm
Do you have Conformer-RNNT results on Librispeech? Do they match the numbers from the Conformer paper? Thanks.
please see https://github.com/wenet-e2e/wenet/tree/main/examples/librispeech/rnnt, the result is far behind the conformer paper. Maybe @yuekaizhang could show more details.
please see https://github.com/wenet-e2e/wenet/tree/main/examples/librispeech/rnnt, the result is far behind the conformer paper. Maybe @yuekaizhang could show more details.
@ruomingp This link's model size is about 34Mb. For mid size model in conformer paper, they got 2.3%, 5.0% on librispeech testsets. I have no idea why we can't reach the paper numbers. Maybe they train more epochs or use other techniques. I was wondering if you have some suggestions.
Thanks for the info! Have you tried Conformer-L?
Thanks for the info! Have you tried Conformer-L?
Not yet. FYI, https://github.com/k2-fsa/icefall/blob/master/egs/librispeech/ASR/RESULTS.md#large this one has similar size with Conformer-L.
Thank your guys for the information.
Hi wenet team,
I want to know more details about the RNN-T result.
Just list the last table here:
My question is: