EmbeddedLLM / vllm-rocm

vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs
https://vllm.readthedocs.io
Apache License 2.0
83 stars 5 forks source link

[Feature]: vllm 0.4.1 in ROCM #27

Open linchen111 opened 2 months ago

linchen111 commented 2 months ago

🚀 The feature, motivation and pitch

Hello, I am using the vllm 0.2.6 image. But when I tried to install a new version of vllm myself, such as 0.4.1, it failed (I was using mi250x). Do you have any plans to update the images in the Docker Hub?

Alternatives

No response

Additional context

No response