baichuan-inc / Baichuan-7B

A large-scale 7B pretraining language model developed by BaiChuan-Inc.
https://huggingface.co/baichuan-inc/baichuan-7B
Apache License 2.0
5.67k stars 506 forks source link

[Question] 参数合并后有什么要注意的吗? 我将7B参数和微调参数合并之后,加载新模型,显存占用超过了24G,这个跟原始7B所需显存差很多?这会是什么导致的 #142

Open Micla-SHL opened 6 months ago

Micla-SHL commented 6 months ago

Required prerequisites

Questions

1

Checklist