Open BoyangGu1 opened 3 months ago
It seems that the saved model is fine, but when combining with the original model, the original model (of course without DOC_SEP
) failed to merge with lora. I don't know how to fix that now.
@danielhanchen
Hmmm will try investigating this
Try resize_model_vocab = 128257 in FastLanguageModel.from_pretrained()
Try resize_model_vocab = 128257 in FastLanguageModel.from_pretrained()
Worked perfectly after applying my base model vocab size with the resize_model_vocab param
I used the following code to sft llama3:
Especially, I added a new special token
DOC_SEP
, then when I try to load the model, it crushed:I got the error:
It seems that the when model is saved, it failed to save with 128257 tokens, but stick with the old 128256 token version (without
DOC_SEP
).Please help!