intel-analytics / ipex-llm

Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, MiniCPM, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, GraphRAG, DeepSpeed, vLLM, FastChat, Axolotl, etc.
Apache License 2.0
6.56k stars 1.25k forks source link

MiniCPM-V-2 run error on Windows MTL iGPU #11408

Open aitss2017 opened 3 months ago

aitss2017 commented 3 months ago

I tried to leverage benchmark tool to test this mutimodal model, meet below error. Model link: https://huggingface.co/openbmb/MiniCPM-V-2 Tool link: https://github.com/intel-analytics/ipex-llm/tree/main/python/llm/dev/benchmark/all-in-one

image

qiuxin2012 commented 3 months ago

Sorry, all-in-one's script doesn't support multimodal LLMs now.