Open reilgun opened 1 year ago
I get the same error. I used the config from llama 2 7b here and only changed path to model. Everything else has been just like the repo here for me.
I've trained qloras successfully with the same docker container in the past. I don't know if it's something about the model I am using that triggers this error. You can see it's spicyboros-2.2
@reilgun For me I found that adding the unk_token is the only thing causing it to fail. Simply commenting out the line seems to work when I use it for LLaMa 2
Any idea what could be causing this error? My configuration is as follows: https://raw.githubusercontent.com/reilgun/runpod-qlora-trainer/master/scripts/run_flash_large.sh