Closed nivibilla closed 6 months ago
Unfortunately, it's a bit difficult to auto-detect exl2 - it's the only quant format that does not ship a quantization config file, so the guesswork won't always work. I requested turboderp to have exl2 quants ship the config too, and it seems he's doing that now. Future exl2 quants shouldn't have this problem, hopefully.
Ah okay np. Thanks for the quick reply!
Your current environment
N/A
🐛 Describe the bug
Loading without specifying
--quantization exl2
tries to load the model with quantisation modeNone
. Manually specifying that it is an exl2 quant works.