Open Mark-DelGrande opened 1 week ago
Oh wait please use https://github.com/unslothai/unsloth/wiki#adding-new-tokens ie
model, tokenizer = FastLanguageModel.from_pretrained(...)
from unsloth import add_new_tokens
add_new_tokens(model, tokenizer, new_tokens = ["<CHARACTER_1>", "<THINKING>", "<SCRATCH_PAD>"])
model = FastLanguageModel.get_peft_model(...)
I am using a chat ML template like this to format prompts:
when I use to resize with this code I got back the response:
Embedding(128258, 4096)
Now I am getting back:
Not sure if this is related but it feels like I t might be an di tired to turn False on and it still gave me back
After fine-tuning Llama 3.1 with the same code my responses went from something like this:
This is what I would like to get out of it and looks like my data I fine tuned on but it has become:
Anyone have any ideas if something changed, and how I can get my end token to be caught again?