-
### 🚀 The feature, motivation and pitch
It is common to have a scenario where folks want to deploy multiple vLLM instances on a single machine due to the machine have several GPUs (commonly 8 GPUs). …
-
### Issue
Hey! *I work at exa*. https://exa.ai
I have been experimenting with Aider for a few days and I think it could benefit from Exa web-search integration.
Exa does semantic search + retu…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a sim…
-
### Your current environment
```text
The output of `python collect_env.py`
```
CODE:
from langchain.llms import VLLM
import time
import uvicorn
app = FastAPI()
llm = VLLM(model="tiiua…
-
Ubuntu 22.04LTS comes with Python 3.10. 22.04LTS is replaced by 24.04LTS. 24.04LTS has Python 3.11. Ubuntu 22.04 with Python 3.10 is causing issues with LLM_Web_search.
```
LLM_Web_search | uber g…
-
### Is there an existing issue for this?
- [X] I have searched the existing issues
### Current Behavior
local LLM dont't work for subdomains and vulns
### Expected Behavior
local LLM
### Steps T…
-
Thanks for adding GraphRAG to RAGbuilder.
I had some questions and suggestions, perhaps you want to chat some time.
* QQ: in graphrag.full_retriever you fetch the vector store data but don't use…
-
### 🚀 The feature, motivation and pitch
In the context of a text-only large language model (LLM), it often truncates the input from the left to ensure the conversation can continue within token limit…
-
### Is there an existing issue for this bug?
- [X] I have searched the existing issues
### 🐛 Describe the bug
Modify the code to adapt to qwen2vl(transformers.Qwen2VLForConditionalGeneration) and f…
-
I have recently being working on swebench where we built distributed eval on top of Modal for faster eval cycles. As a next step, I was hoping to use that setup to execute the patch generated by LLMs …