Please include information about your system, the steps to reproduce the bug, and the version of llama.cpp that you are using. If possible, please provide a minimal code example that reproduces the bug.
I have converted GPT2 into gguf and quantize model into q4_0
Please include information about your system, the steps to reproduce the bug, and the version of llama.cpp that you are using. If possible, please provide a minimal code example that reproduces the bug. I have converted GPT2 into gguf and quantize model into q4_0
but when i run inference i meet this bug
do I make any mistakes? and How can I fix this please?