-
Even with my changes, throttling of LLM calls is still not ideal. Best would be to be able to specify a global rate of tokens/minute, or at least calls / minute and enforce this.
On the other side,…
-
# Trending repositories for C#
1. [**Kareadita / Kavita**](https://github.com/Kareadita/Kavita)
__Kavita is a fast, feature rich, cross platform reading server. Built with a focus…
-
Is it possible to use this for conditional generation?
-
Developers access documentation in many ways: Google, StackOverflow, ChatGPT, local documentation webserver, VScode or terminal (man pages ???). What should we focus on first?
Maybe we should do…
-
# TL;DR
Retrieval mechanisms can be used to provide kind of memory for an LLM.
This memory can be used for semantic search and QA based on retrieved code/documentation fragments.
# Context
…
-
**Is your feature request related to a problem? Please describe.**
the old platform supported string-driven templates.
I'm wondering if you all will be supporting that indefinitely-- mainly becau…
-
### What happened?
I am trying to integrate a locally hosted VLLM server to LiteLLM. The curl request(which works alright) that i am using to hit the vllm server directly is:
```
curl --location 'h…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain.js documentation with the integrated search.
- [X] I used the GitHub search to find a …
-
Emacs version: 29.1 + Spacemacs [openai layer](https://develop.spacemacs.org/layers/+web-services/openai/README.html)
I'm getting this message on startup of emacs each time or when I'm trying to tr…
-
### What happened?
Created a service.
Deployed directly docker from Docker Hub.
Now trying connect via HttpClient:
```
chroma_client = chromadb.HttpClient(host="",
port='443',
ssl…