Open claudiosv opened 11 months ago
vLLM offers an OpenAI compatible HTTP server and faster inference. We'd like to offer it as an option for lmwrapper.
vLLM offers an OpenAI compatible HTTP server and faster inference. We'd like to offer it as an option for lmwrapper.