This is a repository with the data and code for the ACL 2019 paper "When a Good Translation is Wrong in Context: ..." and the EMNLP 2019 paper "Context-Aware Monolingual Repair for Neural Machine Translation"
Hi!
How do I get the model to utilise all available GPU memory on each GPU?
I tried changing --batch-len , --optimizer, --optimizer-opts and some other parameters, but I can't seem to get it to use anything other than 416MiB per GPU.
Hi! How do I get the model to utilise all available GPU memory on each GPU? I tried changing
--batch-len
,--optimizer
,--optimizer-opts
and some other parameters, but I can't seem to get it to use anything other than 416MiB per GPU.Here I'm training 3 models in parallel:
Thanks!