qwopqwop200 / GPTQ-for-LLaMa

4 bits quantization of LLaMA using GPTQ
Apache License 2.0
2.99k stars 459 forks source link

error: block with no terminator, has llvm.cond_br %5624, ^bb2, ^bb3 #283

Open Hukongtao opened 1 year ago

Hukongtao commented 1 year ago

I followed the tutorial in the README to run the code,But when I run this sentence

CUDA_VISIBLE_DEVICES=0 python llama.py ${MODEL_DIR} c4 --wbits 4 --groupsize 128 --load llama7b-4bit-128g.pt --benchmark 2048 --check

I got an error: image