-
### Describe the issue as clearly as possible:
`mlx-community/Phi-3-mini-4k-instruct-4bit` and `mlx-community/Phi-3-mini-4k-instruct-8bit` generate text without spaces when using `outlines`. Interes…
-
I have a intel CPU that supports a number of AVX features, but most of them are not picked up when using ollama. Below is the llama.log file:
system info: AVX = 1 | AVX2 = 0 | AVX512 = 0 | AVX512_…
-
### Describe the bug
The `huggingface-cli` fails to download the `microsoft/phi-3-mini-4k-instruct-onnx` model because the `.incomplete` file of the `.onnx` data file is missing.
I assume the fi…
-
Hi @danemadsen, thanks for your hard work on this!
I'd like to write a userguide in a PR for noobs to figure out which models to use and debug various popular FOSS .gguf models from huggingface. I'…
-
For every model I've downloaded, the speed saturates my bandwidth (~13MB/sec) until it hits 98/99%. Then the download slows to a few tens of KB/s and takes hour(s) to finish.
I've tried multipl…
-
I am unsure of why such a thing occurs. I'm passing in the json.gbnf grammar to restrict the output. I have also set the following parameters -
- seed: 123
- temperature: 0.0
- n_predict: 512
-…
-
![error](https://github.com/user-attachments/assets/c6a351db-0074-4db7-bc68-9b6eb9f3081f)
After running the app.py file and putting the model in the web_app_storage/models folder. I get the this er…
-
While testing phi-3 I have seen a very strange behaviour in MLX that is not present in ollama/llama.cpp.
During inference the first date is systematically wrong (any temperature including 0.0 and an…
-
With many claiming that phi3 mini is uncannily good for it's size, and with larger, actually-useful phi3 models on the way, adding support for this arch is almost certainly worthwhile.
-
### What is the issue?
Hi team im getting this error below -
C:\Windows\System32>ollama run gemma
pulling manifest
Error: Head "https://dd20bb891979d25aebc8bec07b2b3bbc.r2.cloudflarestorage.com/ol…