Open RnMss opened 1 year ago
Hi! Please try to build and run llama.cpp and see if it works.
If it crashes too with similar error, report the problem with llama.cpp
to their repo. They would fix it quicker, since their repo is more popular, and then I can port the fix here.
If it does not crash, we would need to compare the code of llama.cpp
and rwkv.cpp
and guess what can cause the issue.
I tried llama.cpp, and it worked without a crash. Tested on models: opt-1.3b and Chinese-Alpaca-LoRA-13B llama.cpp version: master-53dbba7
I took a look at llama.cpp
version of ggml
. Unfortunately, my and their repo are now too diverged to make sense of any comparisons. Sorry for asking you to test llama.cpp
, I'll stop asking users to do that from now on.
As for the issue, I don't have any ideas how to fix this.
I tried adding compile flags -fcf-protection=none
, which is said to disable the CET instruction set like endbr64
, but it does not help.
It doesn't make sense. I roughly read the code but didn't see anything close to that. The disassembly looks rather real, not like some random data. I'm dooooomed.
@RnMss I've updated ggml
to the latest version. Please try again, don't forget to update git submodules (or better -- clone from scratch git clone --recursive https://github.com/saharNooby/rwkv.cpp.git
).
It still doest not work on my CPU. I'll try on Windows later.
Got the same problem in docker nvcr.io/nvidia/pytorch:23.05-py3, tokenizers-0.13.3
try recompile the repo with disable the AVX instruction flag on cmakelist.txt @RnMss . this step works for me
My build crashes inferencing with a model with "Illegal Instruction". I debugged it and seems to crash on an
endbr64
instruction. I think my CPU doesn't support the instruction set. Is there a building option to turn off the instruction set?Version: Master, commit e84c446d9533dabef2d8d60735d5924db63362ff
Command to reproduce
python rwkv/chat_with_bot.py ../models/xxxxxxx.bin
It crashed with "Illegal Instruction"
I debugged the program: