Open mgoin opened 1 month ago
vLLM is a high-throughput and memory-efficient open-source serving engine for LLMs.
vLLM is fast with:
Transparent Logo:
vLLM is a high-throughput and memory-efficient open-source serving engine for LLMs.
vLLM is fast with:
Transparent Logo:![vllm-logo-text-light](https://github.com/huggingface/huggingface.js/assets/3195154/44a9ab52-8ec6-4d16-8850-895508d0d172)