Open YerongLi opened 1 year ago
same question? also how to apply this for unsupported-models
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
For anyone looking for a way to apply Bitsandbytes quantization to custom model, here's my recent open-source project of implementing a custom QLoRA to fine-tune LLM, with basic tools like PyTorch and Bitsandbytes, completely decoupled from hugging face.
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
I want to quantize model from open-flamingo or https://github.com/open-mmlab/Multimodal-GPT (open-flamingo v1) before lora training,
https://github.com/open-mmlab/Multimodal-GPT this repo provides lora, but how to do qlora training and quantize the model beforehand?
Are there any ways to do this fast?