Open Rorschaaaach opened 1 year ago
模型导出:model = AutoModel.from_pretrained(xxx) model = llm.from_hf(model, tokenizer, dtype = "float16") model.save(xxx)
模型载入:llm.model(xxx)
报错内容:Segmentation fault (core dumped)
是因为微调后导致的嘛?
是用PEFT进行的微调吗
lora微调的,刚刚用tools/scripts/chatglm_export.py重新生成了一遍,可以用了
你是参数PEFT加载之后再 export 吗、 加速效果明显吗、
模型导出:model = AutoModel.from_pretrained(xxx) model = llm.from_hf(model, tokenizer, dtype = "float16") model.save(xxx)
模型载入:llm.model(xxx)
报错内容:Segmentation fault (core dumped)
是因为微调后导致的嘛?