aitsc / GLMKD

Are Intermediate Layers and Labels Really Necessary? A General Language Model Distillation Method ; GKD: A General Knowledge Distillation Framework for Large-scale Pre-trained Language Model
MIT License
31 stars 1 forks source link

train data #5

Open yangyingchun1999 opened 3 months ago

yangyingchun1999 commented 3 months ago

Thank you for sharing your work. I am preparing to replicate your study, and I would like to know how I should obtain the data?

aitsc commented 3 months ago

All training data are from public datasets, which are referenced in the paper.