Open iamkhalidbashir opened 1 month ago
multi-modal models like Qwen2-VL is supported in pods, but it seems not in serverless endpoints. see my comment here: https://github.com/runpod-workers/worker-vllm/issues/114#issuecomment-2412967748
yes, this is fixed in the dev branch of vllm which when released in the next version should be fixed for endpoints as well.
Hi there!, At the moment we do not support visual-LLMs, but we are working on support for these in the future
I am trying to run
https://huggingface.co/Qwen/Qwen2-VL-7B-Instruct
on serverless but it throws error