Closed Daviljoe193 closed 10 months ago
This won't happen with all models, only some that have tokens that aren't control symbols but still can't be decoded by the tokenizer. This was necessary to support Deepseek models and to work around a bug in the Tokenizers library. There were some unintended side effects, but they should already be fixed. The fix just didn't make it into the 0.0.10 release.
I'll make sure to keep an eye out for that release, and reopen the issue if the issue persists in that version.
Pretty much as written in the title. I'm using my own personal Colab notebook to run Oobabooga, and while EXL2 models load just fine, GPTQ models specifically won't load properly, with the below error.
This error doesn't happen if I downgrade to ExLlama-V2 0.0.9, so this is likely some sort of regression between the two versions. My terrible workaround is simply to add
!perl -i -pe 's|v0.0.10/exllamav2-0.0.10|v0.0.9/exllamav2-0.0.9|g' temp_requirements.txt
to one of the cells prior to thepip
stuff. That line needs to be commented out to recreate the issue without the workaround, though presumably this also happens on a local install the same way.