Closed ZhangTP1996 closed 1 year ago
Hey, thanks for your question!
Until now, we only used "distilgpt2" and "gpt2-medium" in our experiments.
But, you should be able to use all the pre-trained models from the HuggingFace Model Hub (https://huggingface.co/models) as large language model (llm) - also "gpt2".
Thanks for the great work! I am a newbie in NLP so this might be a silly question. I am wondering how to use the GPT-2 model for GReaT? Is it
or are there any examples that run GReaT instead of Distill-GReaT?