Open thisiskofi opened 2 months ago
You can use vllm for inference. lmdeploy does not support this; I will update the documentation.
Interesting. I'm able to run pure lmdeploy (no swift) on this model: xtuner/llava-llama-3-8b-v1_1-hf
on Huggingface. I'm curious what the difference is.
Describe the bug
Error:
According to the supported models list, this model should be supported by lmdeploy.