Closed gpxhyfzqz closed 4 months ago
That is a bit wierd. Do you also see such behavior in loss or (reward - reward_baseline)?
Thanks for the reply, This is the first time I've noticed this issue, and my curves were normal before。 This is my training loss. This is CIDEr scores in test。
Your train_loss is a little different from mine.
The loss curve of my other experiment is also different from yours, but his CIDEr score curve looks normal. I'm revisiting my code, looking for a problem. Maybe it's a matter of train loss in SCST. If I find the problem, I'll reply later.
I also have a problem, in my previous experiments, when my batch_size settings were larger, I tended to get much lower CIDEr scores than when the batch_size was smaller. Maybe a bigger batch_size will affect the training? Thank you most sincerely for your answer.
Can you try not using beam search see if that changes? also, please follow https://www.iris.unimore.it/retrieve/95abce70-9754-4d80-8a1e-63962f3803dc/2305.12254.pdf. M2 does not add eos which is wrong.
This is a problem with my code.
UserWarning: Detected call of lr_scheduler.step()
before optimizer.step()
. In PyTorch 1
.1.0 and later, you should call them in the opposite order: optimizer.step()
before lr_scheduler.step()
. Failure to do this will result in PyTorch skipping the first value of th
e learning rate schedule. See more details at https://pytorch.org/docs/stable/optim.html#how-to-adjust-learning-rate.
Because of this warning, I put scheduler.step() here.
Now, I modified it and the curve looks normal.
Thank you for your help!!
When I'm training, I get such a reward curve, is it normal? The code is as follows, and it comes from https://github.com/weimingboya/DFT/blob/main/common/utils/utils.py def train_scst(model, dataloader, optim, cider, text_field, epoch, device = 'cuda', scheduler = None, args=None):
Training with self-critical