Open user799595 opened 3 weeks ago
I think I found the culprit, according to https://huggingface.co/unsloth/llama-3-8b-Instruct-bnb-4bit/blob/main/generation_config.json#L4 , since there are 2 eos tokens the model can either generate token 128009 or 128001.
Weird thing is that according to https://huggingface.co/unsloth/llama-3-8b-Instruct-bnb-4bit/blob/main/special_tokens_map.json#L10 , the eos token should be <|eot_id|>
and not both of them so i'm not sure what to make out of this, i'll continue looking at this at my spare time, but I wanted to report this first.
System Info
transformers: 4.44.1 Platform: Linux-5.15.0 Python: 3.10.6 PyTorch: 2.4.0+cu121
Who can help?
@gante
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
Expected behavior
I would expect the eos_tokens used to be consistent.
With
num_beams = 2
(beam search)With
num_beams = 1
(greedy search)