artidoro / qlora

QLoRA: Efficient Finetuning of Quantized LLMs
https://arxiv.org/abs/2305.14314
MIT License
9.74k stars 800 forks source link

Quantization aware finetuning? #273

Open SinanAkkoyun opened 9 months ago

SinanAkkoyun commented 9 months ago

Hi! Is it possible to finetune with quantization in mind? https://www.tensorflow.org/model_optimization/guide/quantization/training

This way one could hopefully eliminate quantization errors even further