Open olojuwin opened 5 years ago
请问你用knowledge distillation 训练了么?
请问你用knowledge distillation 训练了么?
@jackchinor 没有,
那可能有区别,作者说knowledge distillation 后,性能提升了很多。
那可能有区别,作者说knowledge distillation 后,性能提升了很多。
@jackchinor 我对比的是蒸馏之前的
@jackchinor Do you know how to do knowledge distillation? Thanks.
Do you know how to do knowledge distillation?
sorry,I‘ m also working on kd, still not fully understood.
From my understanding you would run the images through both networks and use distillation loss to update the network for the student; however, I'm not sure if you train the teacher network with the student, pre train it on the same dataset, or use pre trained weights from imagenet and freeze the layers. Does anyone have any idea?
@olojuwin 可以放出训练参数吗
我训练了一个模型,LFW:99.80,CFP: 0.98386, AGEDB: 0.97867,megaface 97左右,但感觉和文章有一些差距