-
### What is the issue?
Hi
I have a little problem, ill try to run the model that i downloaded, but it not start.
I try on many ways:
ollama run qwen2:72b-instruct --verbose
also I try with:…
-
### 🐛 Describe the bug
I try to compile an object detection model that uses torchvison's rpn module. it fails when generating anchors with `AssertionError: Mutating module attribute cell_anchors du…
-
**Describe the Issue**
When sending a specific image to kobold using MiniCPM 2.6 the Kobold server crashes. Works with Llava 1.5.
**Additional Information:**
Image causing crash:
![2024-02-24_17…
-
YellowFin is a novel method for tuning momentum hyperparameters (learning rate & momentum). It blends the nice sides of momentum SGD (convergence properties) and Adam and similar (automatic tuning of …
-
### Prerequisites
- [X] I am running the latest code. Mention the version if possible as well.
- [X] I carefully followed the [README.md](https://github.com/ggerganov/llama.cpp/blob/master/README.…
-
For each publication in neuroelectro, calculate the pairwise similarity of electrode solutions (or some other set of experimental metadata). For these same publications' last authors, calculate the ne…
-
### What is the issue?
This is deployed via Helm with a Radeon VII. I'm aware gfx906 is no longer supported and is in maintenance mode. Let me know if you suggest using a different version of ollama.…
-
Platform (like ubuntu 16.04):
Python version:Python 3.6.3
Source framework with version (like Tensorflow 1.4.1 with GPU): caffe
Destination framework with version (like CNTK 2.3 with GPU): ON…
-
### What happened?
I am trying to run inference on RPC example. When running the llama-cli with rpc feature over a single rpc-server on localhost, the inference throughput is only 1.9 tok/sec for lla…
-
### What happened?
my run u:\llama\llama.cpp\build\bin\llama-cli.exe -mli -co -fa -ngl 64 -cnv --chat-template gemma -m llama3-8B-Chinese-Chat-q8.gguf
win11 amd 7900x hip 6.1 vs 2022
cmake -DGGM…