songmzhang / DSKD

Repo for Paper "Dual-Space Knowledge Distillation for Large Language Models".
29 stars 3 forks source link

Is the Tinyllama in the description a base model or pretrained model? #16

Open survivebycoding opened 2 weeks ago

songmzhang commented 2 weeks ago

They are all base models.