Closed zxbjushuai closed 3 months ago
I want to quantize llama2_7B but OOM.My cuda memory(48GB) is not enough for llama2_7B,so I want to try models that take up less memory
Only a few parameters are changed here in LLAMA-7B.sh, you should change the model path with your OPT-1.3B path!
oh,thank you very much!🤗
I want to quantize llama2_7B but OOM.My cuda memory(48GB) is not enough for llama2_7B,so I want to try models that take up less memory