issues
search
aitsc
/
GLMKD
Are Intermediate Layers and Labels Really Necessary? A General Language Model Distillation Method ; GKD: A General Knowledge Distillation Framework for Large-scale Pre-trained Language Model
MIT License
30
stars
1
forks
source link
issues
Newest
Newest
Most commented
Recently updated
Oldest
Least commented
Least recently updated
Using different language models
#4
j-datta
opened
1 week ago
4
Are there alternatives for downloading the models?
#3
rafiiary
opened
5 months ago
4
Compatibility with chatGLM2 and chatGLM3
#2
Ishiki-Iroha
opened
6 months ago
1
关于您最新的论文中2b模型的效果
#1
Kausal-Lei
closed
11 months ago
2