artidoro / qlora

QLoRA: Efficient Finetuning of Quantized LLMs
https://arxiv.org/abs/2305.14314
MIT License
10.08k stars 823 forks source link

Quantization aware finetuning? #273

Open SinanAkkoyun opened 1 year ago

SinanAkkoyun commented 1 year ago

Hi! Is it possible to finetune with quantization in mind? https://www.tensorflow.org/model_optimization/guide/quantization/training

This way one could hopefully eliminate quantization errors even further