Issue:
I ran into tokenization mismatch errors when I tried to fine-tune from Llama-3.1. I pre-trained a new MLP adapter for Llama-3.1 and that seems to work, but the fine-tuning script produced lots of warnings like the following and the loss was always 0.
WARNING: tokenization mismatch: 384 vs. 392. (ignored)
WARNING: tokenization mismatch: 414 vs. 416. (ignored)
WARNING: tokenization mismatch: 268 vs. 270. (ignored)
WARNING: tokenization mismatch: 273 vs. 275. (ignored)
WARNING: tokenization mismatch: 284 vs. 286. (ignored)
WARNING: tokenization mismatch: 218 vs. 220. (ignored)
Describe the issue
Issue: I ran into tokenization mismatch errors when I tried to fine-tune from Llama-3.1. I pre-trained a new MLP adapter for Llama-3.1 and that seems to work, but the fine-tuning script produced lots of warnings like the following and the loss was always 0.
Command:
Environment: Here are the versions of some important packages in my conda environment.