Closed JinYu1998 closed 1 year ago
--have_mlp 1
and run the script run_cifar_distill.sh
, it can automatically learn the temperature during training. You can obtain the vanilla distillation results without CTKD as long as you set --have_mlp 0
.python train_student.py --path-t ./save/models/resnet56_vanilla/ckpt_epoch_240.pth --distill dkd --model_s resnet20 -r 1 -a 0 -b 1 --dkd_alpha 1 --dkd_beta 2
- The CTKD method has been integrated into other distillation methods. When you set
--have_mlp 1
and run the scriptrun_cifar_distill.sh
, it can automatically learn the temperature during training. You can obtain the vanilla distillation results without CTKD as long as you set--have_mlp 0
.- Sorry about the mistake. Try this :
python train_student.py --path-t ./save/models/resnet56_vanilla/ckpt_epoch_240.pth --distill dkd --model_s resnet20 -r 1 -a 0 -b 1 --dkd_alpha 1 --dkd_beta 2
OK, thanks for your replay. I still have some questions:
--distill kd
?KL Div
).Thanks again for your reply!
Hi~ I will close this issue tomorrow if you have no more questions. Feel free to contact me via email: zhengli97 [at] mail.nankai.edu.cn
okk~~ thanks again for your reply. Have a great life~
Why is there no code for CTKD in the run_cifar_distill.sh file? And the code about 14 lines of DKD seems to be wrong!