vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
29.65k stars 4.47k forks source link

[Usage]: Model architectures ['LlavaNextForConditionalGeneration'] are not supported for now #10065

Open BITDDD opened 2 hours ago

BITDDD commented 2 hours ago

Your current environment

vllm==0.6.1.post1 vllm-flash-attn==2.6.1 image

How would you like to use vllm

I want use LLAVA-1.6-hf-mistral-7B,Tips Model architectures ['LlavaNextForConditionalGeneration'] are not supported for now,but LlavaNextForConditionalGeneration in the Supported architectures.

Before submitting a new issue...

DarkLight1337 commented 2 hours ago

Are there any exception stack traces being printed before this error? Maybe vLLM failed to import the model.

BITDDD commented 2 hours ago

Are there any exception stack traces being printed before this error? Maybe vLLM failed to import the model. image

DarkLight1337 commented 2 hours ago

Hmm... can you try updating your vLLM version and see if the issue persists?

BITDDD commented 2 hours ago

Hmm... can you try updating your vLLM version and see if the issue persists? image I Try pip install -U vllm,now version vllm--0.6.3.post1,the same question occur ...

DarkLight1337 commented 2 hours ago

Are there any exception stack traces being printed before this error? Maybe vLLM failed to import the model. image

Can you show the full logs in this cell? Not just the error.

BITDDD commented 1 hour ago

Are there any exception stack traces being printed before this error? Maybe vLLM failed to import the model. image

Can you show the full logs in this cell? Not just the error.

`from io import BytesIO

import requests from PIL import Image

from vllm import LLM, SamplingParams import os

def run_llava_next():

llm = LLM(
    model="llava-hf/llava-v1.6-mistral-7b-hf",
    image_token_id=32000,
    image_input_shape="1,3,336,336",
    image_feature_size=1176,
)

prompt = "[INST] " + "<image>" * 1176 + (
    "\nWhat is shown in this image? [/INST]")
url = "https://h2o-release.s3.amazonaws.com/h2ogpt/bigben.jpg"
image = Image.open(BytesIO(requests.get(url).content))
sampling_params = SamplingParams(temperature=0.8,
                                 top_p=0.95,
                                 max_tokens=100)

outputs = llm.generate(
    {
        "prompt": prompt,
        "multi_modal_data": {
            "image": image
        }
    },
    sampling_params=sampling_params)

generated_text = ""
for o in outputs:
    generated_text += o.outputs[0].text

print(f"LLM output:{generated_text}")

if name == "main": run_llava_next()` image image My code and full logs

DarkLight1337 commented 1 hour ago

Do you not get any INFO messages in your output? Maybe you have to configure your Jupyter notebook to output those...

BITDDD commented 1 hour ago

Do you not get any INFO messages in your output? Maybe you have to configure your Jupyter notebook to output those...

emmm,It seems that there is no info messages...

DarkLight1337 commented 1 hour ago

Can you try running this code in a Python script instead of notebook?

BITDDD commented 1 hour ago

Can you try running this code in a Python script instead of notebook?

I'll try it out and synchronize new progress here. Thank you for your time.