Closed raotnameh closed 2 years ago
Hi @raotnameh , decoding depends on the beam size, size of language model, etc. Also the current fairseq implementation doesn't provide for batch inference, hence the process is slow (and runs on single-thread)! We are currently working on to release a HF-compatible model. That should speed up the decoding (through batch inferencing).
I am using docker.