vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
27.95k stars 4.13k forks source link

[Usage]: Why compute_full_blocks_in_seq in block manager v1 needs -1 #6288

Closed Juelianqvq closed 1 month ago

Juelianqvq commented 2 months ago

Your current environment

Why in this line needs -1 ? https://github.com/vllm-project/vllm/blob/main/vllm/core/block_manager_v1.py#L667)

How would you like to use vllm

No response

Juelianqvq commented 2 months ago

@rkooo567

yuki252111 commented 2 months ago

because this is block index

Juelianqvq commented 2 months ago

because this is block index

Well, I don't think so.