Closed djwhitt closed 7 months ago
The latest updates on your projects. Learn more about Vercel for Git ↗︎
Name | Status | Preview | Comments | Updated (UTC) |
---|---|---|---|---|
modelfusion | ✅ Ready (Inspect) | Visit Preview | 💬 Add feedback | Dec 10, 2023 5:10pm |
Neat, thank you!
Included in v0.95.0
llama.cpp recently added a cache_prompt parameter to its API: https://github.com/ggerganov/llama.cpp/commit/05cd6e5036d72d0930de4d8f6be7bce09e8dda24
It does the following (described in the llama.cpp server README):
This PR adds support for it. The build runs and the functionality seems to work as intended in my project.