Open hjc3613 opened 2 years ago
same question, have you solved?
I re-tokenize the training data, ensure that punctuation apart from its adjacent word, this resolve the problem. now I face another problem: the ppl and bleu seems too disappointed, which bleu about 8, much worse than 20 generated by opennmt-tf
I had the same problem when using an Apex install of another GPU: A100 - V100.
Therefore my solution was to create seperate conda environments for different GPUs
❓ Questions and Help
Before asking:
What is your question?
when I train an english to chinese model using transformer_iwslt_de_en architecture, after 6 steps, the error occured, which like the follow picture:
What have you tried?
the train args was copyed from fairseq/examples/translation/readme.md, I only add two new args, which is marked as follow:
What's your environment?