Closed sammyf closed 9 months ago
Basically the same as with me and MythoMax (GPTQ) on Ooba. It's specifically ExLlama-V2 0.0.10 that's the problem. It's a known issue, and will be fixed with 0.0.11. #211
Specifically quoting @turboderp
This won't happen with all models, only some that have tokens that aren't control symbols but still can't be decoded by the tokenizer. This was necessary to support Deepseek models and to work around a bug in the Tokenizers library. There were some unintended side effects, but they should already be fixed. The fix just didn't make it into the 0.0.10 release.
Update on this issue. I tried Oobabooga with that same GPTQ quant of MythoMax, but with exllamav2 updated to 0.0.11 (Which was released just an hour ago), and I can confirm that the model now loads correctly, so I'd recommend you retest to see if your model loads fine now on the current version. If it does, then you should be fine closing this issue.
Yeps .I can confirm everything now works again. Thanks for your work.
Caveat : I think it's a problem with exllamav2, but it might be a problem with tabbyAPI
I updated tabbyAPI and ExLllama2 and suddenly older models which used to work return an error with the tokenizer :
load_model with this payload
returns :