ms-swift: Use PEFT or Full-parameter to finetune 250+ LLMs or 35+ MLLMs. (Qwen2, GLM4, Internlm2, Yi, Llama3, Llava, MiniCPM-V, Deepseek, Baichuan2, Phi3-Vision, ...)
2.09k
stars
203
forks
source link
load qwen110B model using get_vllm_engine throws error #1081
Open
phoenixbai opened 2 weeks ago
Describe the bug I try to load qwen110B model using below code for batch inference, but it throws error:
Your hardware and system info
8 A100 gpu cards:
Additional context
error.log
running error logs is as below, detail log is also attached :