vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
25.4k stars 3.67k forks source link

[Feature]: Support for MiniCPM-Llama3-V-2_5 the Multi-modal LLM #4943

Closed wizd closed 1 month ago

wizd commented 3 months ago

🚀 The feature, motivation and pitch

Tested with the latest commit but got error:

[rank0]: ValueError: Model architectures ['MiniCPMV'] are not supported for now. Supported architectures: ['AquilaModel', 'AquilaForCausalLM', 'BaiChuanForCausalLM', 'BaichuanForCausalLM', 'BloomForCausalLM', 'ChatGLMModel', 'ChatGLMForConditionalGeneration', 'CohereForCausalLM', 'DbrxForCausalLM', 'DeciLMForCausalLM', 'DeepseekForCausalLM', 'FalconForCausalLM', 'GemmaForCausalLM', 'GPT2LMHeadModel', 'GPTBigCodeForCausalLM', 'GPTJForCausalLM', 'GPTNeoXForCausalLM', 'InternLMForCausalLM', 'InternLM2ForCausalLM', 'JAISLMHeadModel', 'LlamaForCausalLM', 'LlavaForConditionalGeneration', 'LLaMAForCausalLM', 'MistralForCausalLM', 'MixtralForCausalLM', 'QuantMixtralForCausalLM', 'MptForCausalLM', 'MPTForCausalLM', 'MiniCPMForCausalLM', 'OlmoForCausalLM', 'OPTForCausalLM', 'OrionForCausalLM', 'PhiForCausalLM', 'Phi3ForCausalLM', 'QWenLMHeadModel', 'Qwen2ForCausalLM', 'Qwen2MoeForCausalLM', 'RWForCausalLM', 'StableLMEpochForCausalLM', 'StableLmForCausalLM', 'Starcoder2ForCausalLM', 'ArcticForCausalLM', 'XverseForCausalLM', 'MistralModel']

Alternatives

No response

Additional context

No response

thesby commented 3 months ago

+1

lxjhunan commented 2 months ago

+1

tonhsiao commented 2 months ago

I got the same issue, look forward to hear the good news.

sabarish244 commented 2 months ago

+1 Ya me too facing same issue waiting for the Vllm team to add support for this model

lzjever commented 2 months ago

+1

morestart commented 2 months ago

+1

Brucewuzhang commented 2 months ago

+1

PeterAdam2015 commented 1 month ago

+1

zylo117 commented 1 month ago

+1

DarkLight1337 commented 1 month ago

Closing as completed by #4087.