-
### What is the issue?
`Error: llama runner process has terminated: signal: segmentation fault (core dumped)`. It occurs while loading larger models, that are still within the VRAM capacity. Here I…
-
### What happened?
I am trying to run Ollama there's no specific command I found that run Ollama private model.
There's any further guidance available with current repo?
### Relevant log output
``…
-
**Context**
I use Tabby VSCode extension with a local Tabby server.
Currently, when I start VSCode and the Tabby server is not running, it reminds me of that through the yellow indicated extension i…
-
I am running [2.1.99] on my Macbook Air M2 16GB. Yesterday I wanted to try using nomic-embed-text-v1.5 and then jina-embeddings-v2 instead of bge-micro for embedding. In the case of nomic, the embeddi…
-
### What is the issue?
When I load a large model that doesn't fit in VRAM, Ollama crashes:
➜ ~ ollama run dbrx:132b-instruct-q8_0
Error: llama runner process has terminated: signal: segmentation …
-
If we can tell a model to look at picture we should be able to tell it to read from a text file.
There are so many cases where I want to frame a question with data or text, that just doesn't work.
…
-
We could easily add more models to the list of chat models used for summarization:
https://github.com/enjalot/latent-scope/blob/main/latentscope/models/chat_models.json
There are plenty of small o…
-
simple codes like below
`ollama.chat(model='mistral:instruct', messages=[{'role': 'user', 'content': 'Why is the sky blue?'}])`
OR
```
import ollama
response = ollama.chat(model='mistral:i…
-
# Bug Report
## Description
**Bug Summary:**
When importing and loading a large number of chat sessions into Open WebUI, the application experiences significant performance degradation, causing…
-
Hi there,
Thank you for a wonderful piece of software, it has a very friendly interface and a great take on usability.
I am using ollama with [ollama_proxy_server](https://github.com/ParisNeo/ol…