EmbeddedLLM / vllm

vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs
https://vllm.readthedocs.io
Apache License 2.0
89 stars 5 forks source link

ROCm Port #1

Closed kliuae closed 1 year ago

kliuae commented 1 year ago

Ported to ROCm

Added quick start instructions

Models tested

Co-authored by @tjtanaa @iAmir97 @tanpinsiang @meiyihTan