Open sscheng216 opened 9 months ago
@sscheng216 hi, you may check the type of zeros. You need to transfer the zeros from coded format into fp16
@sscheng216 hi, you may check the type of zeros. You need to transfer the zeros from coded format into fp16
Hi, how to transfer the zeros from coded format into fp16
hi, i also want to ask how to transfer the zeros from coded format into fp16
Hi, I also meet this problem? Anyone solved it? Thank you!
Dear Sir,
Thanks for your sharing of this great work. I followed the instruction in readme.md that using GPTQ-for-LLaMa to quantize the llama-2-7b-chat model and applied lora finetuning, but I got error when using merge.py due to the dimension mismatch of scales and qzeros:
Did I miss something?