vllm-project / vllm

A high-throughput and memory-efficient inference and serving engine for LLMs
https://docs.vllm.ai
Apache License 2.0
29.98k stars 4.53k forks source link

[Usage]: how to acquire logits in vllm #8762

Open zeyang12-jpg opened 1 month ago

zeyang12-jpg commented 1 month ago

Your current environment

The output of `python collect_env.py`

How would you like to use vllm

I want to acquire logits when I run benchmark_throughput.py to do the softmax optimization, but the output in vllm doesn't have logits, how can I acquire it.

Before submitting a new issue...

zeyang12-jpg commented 1 month ago

my env is MI250 and ROCM 6.1

hmellor commented 1 month ago

Would a logits_processor work for your use case? https://docs.vllm.ai/en/latest/dev/sampling_params.html