ModelTC / lightllm

LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalability, and high-speed performance.
Apache License 2.0
2.52k stars 198 forks source link

[BUG] Qwen-7B-Chat AttributeError: 'LlamaSplitFuseInferStateInfo' object has no attribute 'logn_values' #239

Open exceedzhang opened 10 months ago

exceedzhang commented 10 months ago

Before you submit an issue, please search for existing issues to avoid duplicates.

Issue description: AttributeError: 'LlamaSplitFuseInferStateInfo' object has no attribute 'logn_values'

Please provide a clear and concise description of your issue.

Steps to reproduce: python -m lightllm.server.api_server --model_dir /root/autodl-tmp/Qwen-7B-Chat --tp 1 --trust_remote_code --splitfuse_mode

Please list the steps to reproduce the issue, such as:

Expected behavior:

Please describe what you expected to happen.

Error logging:

If applicable, please copy and paste the error message or stack trace here. Use code blocks for better readability.

Environment:

Please provide information about your environment, such as:

Additional context:

image

Language:

lightllm run Qwen-7B-Chat use splitfuse_mode not worked.

hiworldwzj commented 10 months ago

@exceedzhang splitfuse mode is in test. so this mode only support llama and llama2 now.

hiworldwzj commented 10 months ago

we will try to support other mode types soon.

ObliviousDonkey commented 10 months ago

does it support qwen VL?

hiworldwzj commented 10 months ago

@ObliviousDonkey qwen VL will be supported soon.