issues
search
qwopqwop200
/
GPTQ-for-LLaMa
4 bits quantization of LLaMA using GPTQ
Apache License 2.0
2.99k
stars
459
forks
source link
adding missing transformers import to opt.py cuda
#237
Closed
YellowRoseCx
closed
1 year ago