Open sergiogcharles opened 6 months ago
@sergiogcharles , hello!
On local machine I use this configuration in .env (using docker-compose-dev)
But when I run the same on VM into enterprise net I use like this: GEN_AI_API_ENDPOINT=YOUR_VM_IP:11434
Also i run ollama in docker with GPU boost https://ollama.com/blog/ollama-is-now-available-as-an-official-docker-image
Probably this can help you...
@sergiogcharles , hello!
On local machine I use this configuration in .env (using docker-compose-dev)
But when I run the same on VM into enterprise net I use like this: GEN_AI_API_ENDPOINT=YOUR_VM_IP:11434
Also i run ollama in docker with GPU boost https://ollama.com/blog/ollama-is-now-available-as-an-official-docker-image
Probably this can help you...
Hey, thanks for the response. Unfortunately, I am using this exact configuration, but I'm still seeing a network error.
@sergiogcharles I have the same issue with Ollama. Tag v0.3.36 is working fine, so reverting to that is a quickfix. Maybe bumping litellm in v.0.3.37 introduced a bug.
3.36
Hi @chemista, did you have to change anything besides creating the .env config? Also, are you running ollama from within a docker container? Thanks!
@sergiogcharles only .env config. I'm running Ollama via docker on the same machine. Just remember to delete docker volumes from the buggy version before you start the v0.3.36 with docker-compose.
@chemista Still no luck, unfortunately. I'll try some other tags and see if I can get it to work. Thanks for your help!
@sergiogcharles only .env config. I'm running Ollama via docker on the same machine. Just remember to delete docker volumes from the buggy version before you start the v0.3.36 with docker-compose.
@chemista I was curious if you're running Docker with the WSL 2 based engine?
@chemista I got it working! Note to future users: make sure you enable the 'Add the *.docker.internal names' option if you're using the WSL2 based engine. I am using tag 0.3.23.
Hey folks, I've found an issue that is likely causing this for you all. Some arguments which LiteLLM were suppose to drop when not applicable weren't being dropped correctly and was causing Ollama to have some very weird behavior and failing. Hopefully the latest tag works for you all now!
Guide also updated: https://docs.danswer.dev/gen_ai_configs/ollama
Will just mention though that the Ollama model options are still worse than the OpenAI options. If possible, we suggest using GPT4 or GPT4-Turbo. With the guide above, a lot of features are turned off impacting the quality of the experience.
Hello, I have same error failed to send, dropping 1 traces to intake at http://localhost:8126/v0.5/traces after 3 retries, 1 additional messages skipped
with azure open ai (gpt4)
Hey! With the latest version i also get the network error
when using llama2 or llama3. If i just chat with the GPT directly it gives me a response. Just the danswer
and paraphrase
chats are having this issue.
OpenAI works just fine.
Hi, I configured Danswer to use Ollama, running
ollama run llama2
, and verified that the api server responds with curl. It works well in cli, but when I prompt the model in Danswer I seenetwork error
. I looked in the ngnix container logs, but there don't seem to be any errors. I read that this might be a problem with ngnix timeouts, so I tried to change app.conf.template.dev to this, but still no luck:These are the api server's logs:
As you can see, I'm getting
api_server-1 | failed to send, dropping 1 traces to intake at http://localhost:8126/v0.5/traces after 3 retries
.Thanks!