tenstorrent / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
5 stars 1 forks source link

Cache kv blocks for faster initialization, modify model and cache args to allow for higher seq lens #22

Closed skhorasganiTT closed 1 month ago

skhorasganiTT commented 1 month ago