In finetune.py, lines283~284, model.backward(loss) is directly followed by model.step().
Does this mean that the model parameters will immediately update once the loss in the current step is back-propogated and the args.gradient_accumulation_steps is not working? Is this a bug?
In
finetune.py
, lines283~284,model.backward(loss)
is directly followed bymodel.step()
. Does this mean that the model parameters will immediately update once the loss in the current step is back-propogated and theargs.gradient_accumulation_steps
is not working? Is this a bug?