Closed Snnzhao closed 3 years ago
I found even if train 25 epochs when the loss drop to 0.003.
It outputs a lot program like
"Find(farmer)\Relate(occupation\
It sames the split token always wrong
I sames to solve this problem.
This problem is caused by the token \ \
It is due to the error in the latest BART checkpoint. If somebody meets this problem, you can download the previous checkpoint from https://cloud.tsinghua.edu.cn/f/3b59ec6c43034cfc8841/?dl=1 and load it as the pretrained model to achieve our reported results.
In the train.py, the model is not valid during training. So I copy the valid part from predict.py. But during training at start, the valid acc drop from 0.007 to 0.004 and does not increase during training. I use bart-base as pretrained model. Can you provide more details for the experiment?