issues
search
IST-DASLab
/
gptq
Code for the ICLR 2023 paper "GPTQ: Accurate Post-training Quantization of Generative Pretrained Transformers".
https://arxiv.org/abs/2210.17323
Apache License 2.0
1.81k
stars
145
forks
source link
How can we use this lib to quantize Falcon7b / 40b models?
#36
Open
tarunmcom
opened
1 year ago
tarunmcom
commented
1 year ago
How can we use this lib to quantize Falcon7b / 40b models?
How can we use this lib to quantize Falcon7b / 40b models?