Closed robinsonmhj closed 5 months ago
cc @SunMarc
Hi @robinsonmhj, bnb only is not compatible with deepspeeed. In fact, you can't train a quantized model in general. However, it works with peft + bnb + deepspeed (stage 1 and 2). For more detail, please check the following PR.
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
Please note that issues that do not follow the contributing guidelines are likely to be ignored.
Hi @robinsonmhj, bnb only is not compatible with deepspeeed. In fact, you can't train a quantized model in general. However, it works with peft + bnb + deepspeed (stage 1 and 2). For more detail, please check the following https://github.com/huggingface/peft/pull/1529.
@SunMarc sorry to bring up a dead issue here. By this comment (and the attached PR) is it not possible to load a pretrained model using 4bq from bnb, then kick off training using deepspeed?
I keep getting ValueError: .to is not supported for 4-bit or 8-bit bitsandbytes models. Please use the model as it is, since the model has already been set to the correct devices and casted to the correct
type`
I am trying to
System Info
transformers==4.31.0 accelerate==0.21.0 deepspeed==0.13.2 bitsandbytes==0.42.0
Who can help?
No response
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
Expected behavior
expected no error and exception, however, get the following error