-
hello,
I am a student studying knowledge distillation in semantic segmentation.
I want to study your method with implementation detail.
Is it possible to upload your segmentation task code?
-
* Paper: Knowledge distillation: A good teacher is patient and consistent
* Paper Link: https://arxiv.org/abs/2106.05237
### Description
- Paper focuses on 2 important aspects of Knowledge Distil…
-
## 論文リンク
https://arxiv.org/pdf/1606.07947.pdf
## 概要
知識蒸留をトークンレベル、文レベルでNMTに適用
パラメータを少なくして学習の高速化やBLEUスコアの向上を確認している。
## 先行研究との差異
## 手法のキモ
NMTに知識蒸留を適用
毎トークンごとの語彙数次元の分類タスクに適用されることにな…
-
Thank you for your valuable work. I don't know if I can ask u the difference between this repo and https://github.com/CityChan/Federated-Hyper-Knowledge-Distillation, seems this two repo has same cont…
-
Hi,
My name is Marco Mistretta, and I am a researcher at [MICC (Florence, Italy)](https://github.com/miccunifi).
Thank you for creating this repository! It has been an invaluable resource for me.
…
-
Hi, thanks for your sharing.
When I tried to use multi-gpu to train Knowledge Distillation:
`python3 -m torch.distributed.run --nproc_per_node $N_GPU distillation.py ...`
I got the error:
torch.di…
-
When training student generation 2, you use the student weight from generation 1. Isn't that just continue training? Are you resetting the learning rate, or resetting the weight of some part of studen…
-
## 一言でいうと
構造が異なるモデル間で蒸留を行うことで、特定モデルで学習しやすい知識(CNNなら局所特徴、RNNなら系列構造など)を転移できるか検証した研究。CNN=>MLP、LSTM=>Transformerで蒸留を行いそれぞれの学習傾向が蒸留先モデルに反映されることを確認。
### 論文リンク
https://arxiv.org/abs/2006.00555
###…
-
Dear Dr. Thomas Müller:
I have read your papper, the moudel you proposed is really concise and efficient, so I want to used it in my own project. When I read the code in this project I didn't find …
-
I hope this email finds you well. I would like to express my gratitude for your captivating and outstanding work titled "Fine-Tuning Global Model via Data-Free Knowledge Distillation for Non-IID Feder…