-
The insert process is quite slow for a small document. I tried to change `llm_model_max_async` value but the speed is never change. I also saw that the insert process is only using single core of my C…
-
### What is the issue?
I have 8 AMD 7900XTX cards in llama.cpp to limit access to certain GPUs, I use the HIP_VISIBLE_DEVICES command and it works correctly. However, if I want to limit GPU access fo…
-
### What happened?
**Background:**
I'm installing Quivr locally on Ubuntu and I want to use llama3.1 in Ollama.
I changed the URL in .env file using the internal IP address of the host machine, "…
-
### What is the issue?
launchctl setenv OLLAMA_HOST "0.0.0.0" does not work in version 0.3.14. It still listens to 127.0.0.1
### OS
macOS
### GPU
Apple
### CPU
Apple
### Ollama version
0.3.14
-
### First Check
- [X] This is not a feature request.
- [X] I added a very descriptive title to this issue (title field is above this).
- [X] I used the GitHub search to find a similar issue and d…
-
How to use local LMM in this repo?
-
### Before submitting your bug report
- [X] I believe this is a bug. I'll try to join the [Continue Discord](https://discord.gg/NWtdYexhMs) for questions
- [X] I'm not able to find an [open issue]…
-
Hi. Any way to use Ollama with org-ai? I have done this for `chatgpt-shell` here: https://github.com/xenodium/chatgpt-shell/issues/201.
-
I couldn't find anything. But it seems to me that it should be possible to set an API key to access the ollam service. Request header "Authorization: Bearer {token}".
-
**Describe the question**
The same question yields different answers in the API and the UI.
**To Reproduce**
Steps to reproduce the behavior:
The UI
_Question (as shown in the blue box below)…