-
### Plugin Description
Plugin to rename functions and variables with a locally hosted ollama server. Can use any of the openly available models such as llama3 and gemma2.
### Plugin Repository URL
…
-
### What do you need?
After some experimentation, q8 llama3 and mistral **run locally** are performing **much** better after adjusting the prompt order. I propose restructuring the patterns so tha…
-
When doing inference on Gemma-2-2B with Flash Attention 2, I get the following error. It works just fine with Flash Attention disabled.
transformers==4.44.0
torch==2.4.0
flash-attn==2.6.3
python…
-
### System Info
- `transformers` version: 4.42.4
- Platform: macOS-13.4-arm64-arm-64bit
- Python version: 3.10.13
- Huggingface_hub version: 0.23.4
- Safetensors version: 0.4.3
- Accelerate vers…
-
### What is the issue?
So i am running llms with ollama on linux. I downloaded nvtop to check NVIDIA GPU usage. It seems that during inferencing CPU as well GPU usage shoots up. I am not sure why the…
-
### What is the issue?
I have installed the latest stable version (v0.2.1) from [the releases page](https://github.com/ollama/ollama/releases). However, after checking the version, it makes a distinc…
-
### Describe the bug
06:49:32,683 graphrag.index.reporting.file_workflow_callbacks INFO Error executing verb "cluster_graph" in create_base_entity_graph: EmptyNetworkError details=None
06:49:32,68…
-
### What is the issue?
Previously if I interrupted a pull and started it again (Usually after the speed dropped or wouldn't go to completion after 95%), it was able to resume.
However today whe…
-
### When I run the following script
```
import torch
from accelerate import Accelerator, PartialState
from peft import LoraConfig
from tqdm import tqdm
from transformers import AutoTokenizer, …
-
### What is the issue?
I had custom location of ollama model in 2.8 but after 3.0 update it is ignored.
```
ollama run gemma2:27b
pulling manifest
pulling d7e4b00a7d7a... 1% ▕ …