vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
29.92k stars 4.52k forks source link

[New Model]: Request to integrate Chexagent Multimodel in vLLM #7863

Open MohnishJain opened 2 months ago

MohnishJain commented 2 months ago

The model to consider.

https://huggingface.co/StanfordAIMI/CheXagent-8b

The closest model vllm already supports.

https://huggingface.co/microsoft/llava-med-v1.5-mistral-7b

What's your difficulty of supporting the model you want?

when i tried to host chexagent it says, model architecture not supported by vllm.its has integrated qformer inside this.any help in integrating this will be helpful.

Before submitting a new issue...

DarkLight1337 commented 2 months ago

Sorry for the late response, currently BLIP-2 (w/ Q-Former) is supported in vLLM so you may be able to reuse some of the existing code for this model.