Closed cpsu00 closed 2 months ago
For this model, we directly use the checkpoint released by Gu et al. in MiniLLM. The link in readme points to the repo of MiniLLM.
You can download the gpt2 models via the provided commands. Then untar the gpt2.tar
and you will find it in gpt2/train/sft/gpt2-xlarge
.
It works! Thanks for your fast reply.
Hello, I'm not sure where is the GPT2-1.5B pretrained teacher on dolly is located in the link in readme. Can you guide me where to find it?