EmbeddedLLM / vllm

vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs
https://vllm.readthedocs.io
Apache License 2.0
89 stars 5 forks source link

Merge latest working rocm branch #17

Closed tjtanaa closed 11 months ago

tjtanaa commented 11 months ago

Features