Closed Potomac closed 3 years ago
This is weird, as the script is tested for evaluation. What does your dev.txt
file look like?
Dev.txt contains text in english, one sentence by line. The PC I use has 2 graphic cards, so run_clm.py uses the 2 cards for the training, perhaps the bug occurs only when 2 or more graphic card are used for the training ?
The script is tested on 2 GPUs as well as one. Are you sure this file contains enough text to have a least one batch during evaluation? This is the only thing I can think of for not having an eval_loss returned.
The dev.txt file contains 46 lines, the train file contains 268263 lines.
the specifications of the PC I use :
Like I said, the dev file is maybe too short to provide at least one batch and return a loss. You should try with a longer dev file.
This issue has been automatically marked as stale and been closed because it has not had recent activity. Thank you for your contributions.
If you think this still needs to be addressed please comment on this thread.
Environment info
transformers
version: 4.0.0-rc-1Who can help
albert, bert, GPT2, XLM: @LysandreJik Trainer: @sgugger
Information
Model I am using (Bert, XLNet ...): GPT2
The problem arises when using:
To reproduce
Steps to reproduce the behavior:
when I try to print the content of eval_output then there is just one key : "epoch"
the way I execute run_clm.py :
Expected behavior
The evaluation step should run without problems.