-
### What is the issue?
I just setup Ollama on a fresh machine and am running into an issue starting Ollama on larger models.
I am running Ubuntu 22.04.4 LTS with 2 Nvidia Tesla P40 GPUs with Drive…
-
### What is the issue?
Running Mistral 7B instruct, simple prompts take tens of minutes. Task manager shows CPU is in heavy use and GPU is doing nothing. I can run it with quantization normally wit…
-
Hi,
when running ollama, it hangs up after a few times calling "generate".
It shows no error or something, justt hangs up for hours until it is killed manually.
Stopping and then restarting ollama…
-
### What happened?
Offloading 31 layers out of the 33 with an 8b model produces correct results, with 32 layers, the response is incoherent.
33 or more offloaded layers cause the instruction to be…
8XXD8 updated
1 month ago
-
I am trying to write a Rails PG Apache AGE adapter (https://age.apache.org/) a PG GraphDB extension. I have the basics working in development mode, however, I am not sure how to get it to work with b…
-
Hi, thanks for sharing the work. We have performed a rigorous evaluation for the framework on the cifar-10 setting with this code base, for each setting we run the experiments independently 8 times an…
-
### What is the issue?
When I try `ollama run llama3.1:70b`, occur error `Error: llama runner process has terminated: error loading model: unable to allocate backend buffer`
```
C:\Users\sol>olla…
-
i tried to run your program in google colab withouth changing the parameter. and changing some of the lines but mostly regarding display like parser info, etc.
But the result when using gw datasets f…
-
AdaptiveMaxPool3D 和 AdaptiveMaxPool3D 操作不支持
报错信息
```
There's 1 ops are not supported yet
=========== max_pool2d_with_index ===========
```
版本
```
PaddlePaddle-gpu: 2.0.1.post110
Paddle2Onnx: …
-
I'm eager to explore the new Windows ROCm compatibility feature, but I'm encountering an issue with forcing the GFX version. Currently, I'm using the 0.1.29 pre-release.
My setup includes an RX 660…