Closed toxic0berliner closed 2 months ago
Also tailing the ollama logs I see no request coming. During the install in the GUI it was able to request and list the available models on my ollama instance...
Also tried to remove variables from the prompt and test the prompt with my GUI the answer is faster than 90secs by a long shot. My only setting is to increase the context length to 9000 as I have many devices that I'd wish to control...
Did you set Ollama to listen on all interfaces? By default it only allows connections from the local machine.
I think so it's the OLLAMA_HOST=0.0.0.0:11434
env var I added in the ollama systemd service, seems to work since the API is answering me Ollama is running
.
I tried with very simple prompt just in case my ollama was dead slow, I increased the timeout to 180s... Nothing works so far and I don't know where to look next...
Does the computer you are running ollama on show any CPU or GPU usage when you submit a request? Also how many entities are exposed to the voice assistant?
Asking the AI with the local gui does get me an answer and CPU usage is low but not null, journalctl -f -u ollama
does show activity.
I had 200+ entities exposed by ha, I brought it down to 120. But asking ha using the assistant setup with llama_conversation I see no CPU usage and nothing in the logs of ollama :(
It really seems it's not sending anything to ollama but the setup for example is able to list the available models at least so networking looks fine...
Can you set the request timeout in the integration to at least 180 seconds? (double the default) Everything I can find indicates that Ollama is taking too long to respond because you have so many entities exposed.
Yes! Found the issue! I had "use https" ticked but in fact I would need a reverse proxy to add https or I'd need to pass something to ollama to expose https... Disabling it, I saw some activity when I ask assist a question. At first it did timeout, increasing the timeout I now get responses that are all "turned on the lights, error calling {service:lights_turn-on, room=living_room} or similar, but this issue is closed ❤️ Sadly running an LLM on my AMD 5700U without dedicated GPU is real slow yes and I do have many entities to expose so won't be a useful tool for me yet, but thanks for the help and amazing tool ❤️
Describe the bug
![Screenshot_20240427-122323_Home_Assistant](https://github.com/acon96/home-llm/assets/10488124/fa66c9f2-e094-4e02-9d3a-9b0d622c1d00)
Context
I have a PvE CT running ollama on debian. Works fine using https://github.com/open-webui/open-webui both with Mistral and fixt/home-3b-v3:latest The firewall is set to allow all necessary traffic, I can curl http:10.0.10.61:11434 from within the home assistant container without issue it tells me ollama is running. Also tried the same with ollama.lan and the proper DNS record. Can't figure out what's going on...
Logs
If applicable, please upload any error or debug logs output by Home Assistant.