Open Ian321 opened 1 month ago
server.create_chat_completion would not close the exit_stack if an exception happened while calling llama.create_chat_completion.
server.create_chat_completion
llama.create_chat_completion
This PR fixes #1759 and is a superset of #1795.
+1 to this fix - current we are blocked in bumping our version of llama-cpp-python due to this bug
server.create_chat_completion
would not close the exit_stack if an exception happened while callingllama.create_chat_completion
.This PR fixes #1759 and is a superset of #1795.