tientrandinh / Revisiting-Reverse-Distillation

(CVPR 2023) Revisiting Reverse Distillation for Anomaly Detection
MIT License
116 stars 25 forks source link

Is it something wrong with grad acc? #12

Closed ArlixLin closed 9 months ago

ArlixLin commented 9 months ago

I saw that: optimizer_proj.zero_grad() optimizer_distill.zero_grad() loss.backward() if (i + 1) % accumulation_steps == 0: optimizer_proj.step() optimizer_distill.step() did you mean loss.backward() if (i + 1) % accumulation_steps == 0: optimizer_proj.step() optimizer_distill.step() optimizer_proj.zero_grad() optimizer_distill.zero_grad()

tientrandinh commented 9 months ago

Hi, you are right @ArlixLin, I fixed it. Thank you