Closed Andrei-Aksionov closed 1 year ago
You'll need to re-run the prepare_alpaca.py
script. GptNeox based models use a different tokenizer than LLaMA based models. The error is due to a different vocabulary size
Ah yes, this is super frustrating to debug, I also bumped into this @Andrei-Aksionov . We need better errors for that (#428)
You'll need to re-run the prepare_alpaca.py script. GptNeox based models use a different tokenizer than LLaMA based models. The error is due to a different vocabulary size
Yep, that was exactly the problem.
Duplicate of #366
Hi there 👋
I've noticed a weird behavior when tried to fine-tune OpenLLaMA-3B model:
and got this error output:
what't weird is that other models like
Pythia
orStableLM
works fine. Seems like something related to a PyTorch but don't know why only this model is affected (though I tested only Pythia, StableLM and OpenLLaMA).