google-research / distilling-step-by-step

Apache License 2.0
397 stars 55 forks source link

Distilling failed #7

Open Yanwanxi opened 11 months ago

Yanwanxi commented 11 months ago

Hello, I try to run and train python run.py --from_pretrained google/t5-v1_1-base --dataset cqa --model_type task_prefix --label_type llm --llm palm --alpha 0.5 --batch_size 64 The google/t5-v1_1-base model was downloaded from Hugging Face, but the tokenizer was having problems. 7633a7b7-bd47-4528-b6cd-c5d2d01c695f

rohnson1999 commented 9 months ago

you can manually download files on Hugging Face and set offline mode to run training :) python run.py --from_pretrained '/root/.cache/huggingface/hub/models--google--t5-v1_1-base' --dataset cqa --model_type task_prefix --label_type gt --llm palm --alpha 0.5 --batch_size 64