Closed srikhetramohanty closed 1 week ago
Hi @songmzhang I tried running the evaluation shell script with TinyLlama and it threw the error :
Any leads into what might be the issue here? Mistral checkpoint ran fine with this modified evaluation shell script.
Note : MODEL_PATH="/---/---/DSKD/model_hub/tinyllama/tinyllama-1.1b-3T"
Terminal command : bash scripts/tinyllama/eval_sample.sh $”/---/---/DSKD/outputs/tinyllama/tinyllama-1.1b-3T/sft/criterion=cross_entropy__lora-rank=256-alpha=8-dropout=0.1-bf16__epoch=10__bsz=8x1x1=8__lr=0.001/epoch10_step14300_loss2.2730_rougel27.6927” $16
bash scripts/tinyllama/eval_sample.sh $”/---/---/DSKD/outputs/tinyllama/tinyllama-1.1b-3T/sft/criterion=cross_entropy__lora-rank=256-alpha=8-dropout=0.1-bf16__epoch=10__bsz=8x1x1=8__lr=0.001/epoch10_step14300_loss2.2730_rougel27.6927” $16
Hi, this is because we missed adding the tokenizer class of TinyLLama to the def get_tokenizer() in utils.py. Now we have fixed this bug and you can pull the latest code version.
def get_tokenizer()
utils.py
Thanks. It worked.
Hi @songmzhang I tried running the evaluation shell script with TinyLlama and it threw the error :
Any leads into what might be the issue here? Mistral checkpoint ran fine with this modified evaluation shell script.
Note : MODEL_PATH="/---/---/DSKD/model_hub/tinyllama/tinyllama-1.1b-3T"
Terminal command :
bash scripts/tinyllama/eval_sample.sh $”/---/---/DSKD/outputs/tinyllama/tinyllama-1.1b-3T/sft/criterion=cross_entropy__lora-rank=256-alpha=8-dropout=0.1-bf16__epoch=10__bsz=8x1x1=8__lr=0.001/epoch10_step14300_loss2.2730_rougel27.6927” $16