Closed nguyenhoangthuan99 closed 1 week ago
✅ QA: on cortex.llama-cpp v0.1.37-01.11.24
API request: v1/chat/completions
"stream_options": {
"include_usage": false | true
},
include_usage = false
:include_usage = true
: additional chunk before [DONE] showing token usage statistics
Problem
stream_options
reference: https://platform.openai.com/docs/api-reference/chat/create#chat-create-stream_options
related issue: https://github.com/janhq/internal/issues/160