Closed fzyzcjy closed 1 month ago
If this is on Colab / Kaggle, it's mainly because safetensors are slower to do :(
See https://github.com/unslothai/unsloth/wiki#saving-to-safetensors-not-bin-format-in-colab - ie set safe_serialization = None
I see, thank you!
Error
If use workaround in #278
it can be saved, but cannot be loaded by vllm:
Related issues
Update
vllm now works after upgrading that. However, I would appreciate it if I could save the safetensors format!