How can I pass the system_prompt here? It's a very long text (~2000 characters). We can pass it with each API request (4497)!, but I feel that's not very efficient!
So, I'm thinking of adding this in the startup command itself, but I can't find it in the docs. I prefer adding it through a file, like --system_prompt=system_prompt.txt. Is something like this possible?
Before submitting a new issue...
[X] Make sure you already searched for relevant issues, and asked the chatbot living at the bottom right corner of the documentation page, which can answer lots of frequently asked questions.
Anything you want to discuss about vllm.
So, I run the model using the following Docker command:
How can I pass the system_prompt here? It's a very long text (~2000 characters). We can pass it with each API request (4497)!, but I feel that's not very efficient!
So, I'm thinking of adding this in the startup command itself, but I can't find it in the docs. I prefer adding it through a file, like --system_prompt=system_prompt.txt. Is something like this possible?
Before submitting a new issue...