Closed kenny019 closed 10 months ago
feat: add quantization environment option fix: set lower default max_num_batched_tokens
Resolves Update for vllm 0.2.0 #9
Thank you for your work @kenny019, main branch is now running vllm 0.2.1.post1. Commit: 4f792062aaea02c526ee906979925b447811ef48
feat: add quantization environment option fix: set lower default max_num_batched_tokens
Resolves Update for vllm 0.2.0 #9