Closed konradkalita closed 3 days ago
Thanks for the report. This happens because the config still have the quantization_config
. Would you like to open a PR to fix it ? I think all you need to do is to add in the dequantize()
method the following:
del self.config.quantization_config
System Info
transformers
version: 4.46.0bitsandbytes
version: 0.44.1Who can help?
quantization: @SunMarc @MekkCyber
Information
Tasks
examples
folder (such as GLUE/SQuAD, ...)Reproduction
Stacktrace:
Expected behavior
Dequantized model can be saved using
save_pretrained
and properly loaded usingfrom_pretrained