-
While testing phi-3 I have seen a very strange behaviour in MLX that is not present in ollama/llama.cpp.
During inference the first date is systematically wrong (any temperature including 0.0 and an…
-
### What is the issue?
Hi team im getting this error below -
C:\Windows\System32>ollama run gemma
pulling manifest
Error: Head "https://dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com/ol…
-
### What is the issue?
Ollama is failing to run on GPU instead it uses CPU. If I force it using `HSA_OVERRIDE_GFX_VERSION=9.0.0` then I get `Error: llama runner process has terminated: signal: abo…
-
With many claiming that phi3 mini is uncannily good for it's size, and with larger, actually-useful phi3 models on the way, adding support for this arch is almost certainly worthwhile.
-
**Is your feature request related to a problem? Please describe.**
My company unfortunately cannot benefit from the AI power-tools because anything that involves code remote upload will be violatin…
-
### What is the issue?
We are setting OLLAMA_MAX_LOADED_MODELS=4 in our systemd override file for the ollama service:
![image](https://github.com/ollama/ollama/assets/48829375/b09c1dda-a196-4b89-b34…
-
![error](https://github.com/user-attachments/assets/c6a351db-0074-4db7-bc68-9b6eb9f3081f)
After running the app.py file and putting the model in the web_app_storage/models folder. I get the this er…
-
how to use onnx model for Phi-3 mini 128k for faster inference for local machine having cpu only. Can you provide the code to do it.
-
Is it possible to create a memgpt feature and make it available to all the agents rather than having a separate agent like it's discussed in #530?
-
While rare, `ollama pull` will sometimes result in a digest mismatch on download
```
% ollama run wizard-vicuna-uncensored:30b-q5_K_M
pulling manifest
pulling b1571c5cbd28... 100% |█████████████…