Open ggerganov opened 2 months ago
Not a REST API breaking change, but is server-related: some environment variables are changed in https://github.com/ggerganov/llama.cpp/pull/9308
After #9398, in the completion response seed
contains the seed requested by the user, while seed_cur
contains the seed used to generate the completion. The values can be different if seed
is LLAMA_DEFAULT_SEED
(or -1), in which case a random seed is generated and returned in seed_cur
.
Breaking change #9776 : better security control for public deployments
/slots
endpoint is now disabled by default, start server with --slots
to enable it/slots
and /props
) requires a correct API key to access./health
and /models
are always publicly accessible"system_prompt"
is removed from /completions
endpoint. It is now moved to POST /props
(see documentation)Please note that GET /props
is always enabled to avoid breaking the web UI.
Breaking change for /slots
endpoint https://github.com/ggerganov/llama.cpp/pull/10162
slot[i].state
is removed and replaced by slot[i].is_processing
slot[i].is_processing === false
means the slot is idle
Breaking change for
/slots
endpoint #10162
slot[i].state
is removed and replaced byslot[i].is_processing
slot[i].is_processing === false
means the slot is idle
Was the slots
endpoint also disabled by default? (or maybe just a documentation change?)
https://github.com/ggerganov/llama.cpp/pull/10162/files#diff-42ce5869652f266b01a5b5bc95f4d945db304ce54545e2d0c017886a7f1cee1aR698
For security reasons, "/slots" was disabled by default since https://github.com/ggerganov/llama.cpp/pull/9776 , and was mentioned in the breaking changes table. I just forgot to update the docs.
Not an API change, but maybe good to know that the default web UI for llama-server
changed in https://github.com/ggerganov/llama.cpp/pull/10175
If you want to use the old completion UI, please follow instruction in the PR.
cache_prompt: true
is now used by default (#10501)
Overview
This is a list of changes to the public HTTP interface of the
llama-server
example. Collaborators are encouraged to edit this post in order to reflect important changes to the API that end up merged into themaster
branch.If you are building a 3rd party project that relies on
llama-server
, it is recommended to follow this issue and check it carefully before upgrading to new versions.See also:
libllama
APIRecent API changes (most recent at the top)
/slots
endpoint: removeslot[i].state
, addslot[i].is_processing
/slots
is now disabled by defaultEndpoints now check for API key if it's set
/rerank
endpoint[DONE]\n\n
in OAI stream response to match specseed_cur
to completion response/health
and/slots
For older changes, use:
Upcoming API changes