-
Hi, the normal chat completion seems to work; however, whenever I call the streaming parameter it fails.
I am using VLLM verision 0.2.7
First I start,
`python -m vllm.entrypoints.openai.api_serv…
-
### System Info
```console
INFO text_generation_launcher: Runtime environment:
Target: x86_64-unknown-linux-gnu
Cargo version: 1.75.0
Commit sha: 4ee0a0c4010b6e000f176977648aa1749339e8cb
Doc…
Smana updated
5 months ago
-
### Bug Description
It's literally impossible to fix this issue without intervention from LLAMA INDEX team
Please upgrade it's dep. -> llama-index-vector-stores-chroma 0.0.1 depends on llama-i…
-
### Environment
Operating System: Macbook Pro M1
Python Version: 3.11
Description
I'm encountering an issue when running the setup script for my project. The script is supposed to download an em…
-
- using WSL
- running vanilla ollama with default config, no issues with ollama
- pyenv python 3.11.9 installed and running with Torch, TensorFlow, Flax, and PyTorch added
- all install steps follo…
-
Below is the docker log for your reference.
2024-04-22 16:53:57 11:23:57.071 [INFO ] private_gpt.settings.settings_loader - Starting application with profiles=['default', 'docker']
2024-04-22 16:…
-
Hello,
I've installed privateGPT with Pyenv and Poetry on my MacBook M2 to set up a local RAG using LM Studio version 0.2.21.
I'm using the settings-vllm.yaml configuration file with the followi…
-
### Checked other resources
- [X] I added a very descriptive title to this issue.
- [X] I searched the LangChain documentation with the integrated search.
- [X] I used the GitHub search to find a sim…
-
### Your current environment
```text
PyTorch version: 2.3.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Manjaro Linux (x86_64)
GCC version: …
-
### System Info
serverless inference endpoints
### Information
- [ ] Docker
- [X] The CLI directly
### Tasks
- [X] An officially supported command
- [ ] My own modifications
### Reproduction
Qu…