InternLM / lmdeploy

LMDeploy is a toolkit for compressing, deploying, and serving LLMs.
https://lmdeploy.readthedocs.io/en/latest/
Apache License 2.0
4.09k stars 372 forks source link

[Feature] Support logprob in VLM api server #2041

Open cjfcsjt opened 1 month ago

cjfcsjt commented 1 month ago

Motivation

I found that the input token logprob is supported by Offline Inference Pipeline, as mentioned in doc. However, when I deployed a VLM server, I found that this feature was not supported. How to enable this on the server? Thanks.

Related resources

The code I deploy a server is lmdeploy serve api_server openbmb/MiniCPM-Llama3-V-2_5 --server-port 23333 --session-len 4096

Additional context

No response

cjfcsjt commented 1 month ago

Anyone can help?