issues
search
songmzhang
/
DSKD
Repo for Paper "Dual-Space Knowledge Distillation for Large Language Models".
29
stars
3
forks
source link
Is the Tinyllama in the description a base model or pretrained model?
#16
Open
survivebycoding
opened
2 weeks ago
songmzhang
commented
2 weeks ago
They are all base models.
They are all base models.