Closed win10ogod closed 3 months ago
@LostRuins Can you please update llama.cpp? The strange thing is that even mmap can't detect it. ftype can’t even be recognized.
Hello, can you please try the latest release and see if it works for you now?
We can probably check this one off. Support should be stable for it now.
The model quantified using the latest version of llmam.cpp cannot be used in koboldcpp. The Q4-0 model is identified as IQ3_XXS - 3.0625 bpw. After turning mmap off, the error generated is different.
If mmap is not turned off, the error generated is different. It cannot run with or without BLAS.