hiyouga / LLaMA-Factory

Unify Efficient Fine-Tuning of 100+ LLMs
Apache License 2.0
25.52k stars 3.16k forks source link

910b qwen2 lora生成的模型如何合并权重 #4632

Closed wphtrying closed 2 days ago

wphtrying commented 2 days ago

Reminder

System Info

可以使用这个yaml吗。export_device以及template应该填什么

Note: DO NOT use quantized model or quantization_bit when merging lora adapters

model

model_name_or_path: meta-llama/Meta-Llama-3-8B-Instruct adapter_name_or_path: saves/llama3-8b/lora/sft template: llama3 finetuning_type: lora

export

export_dir: models/llama3_lora_sft export_size: 2 export_device: cpu export_legacy_format: false

Reproduction

CUDA_VISIBLE_DEVICES=0 llamafactory-cli export examples/merge_lora/llama3_lora_sft.yaml

Expected behavior

No response

Others

No response

hiyouga commented 2 days ago

可以