-
### Please check that this issue hasn't been reported before.
- [X] I searched previous [Bug Reports](https://github.com/axolotl-ai-cloud/axolotl/labels/bug) didn't find any similar reports.
###…
-
I am performing a Mega Merge using LLaMA 3.2 3B, both the base model and fine-tuning/instruction tuning, with the DARE linear method. Following the successful completion of the initial merge, I encoun…
-
### System Info
- `transformers` version: 4.44.2
- Platform: Linux-5.10.220-209.869.amzn2.x86_64-x86_64-with-glibc2.26
- Python version: 3.10.14
- Huggingface_hub version: 0.25.1
- Safetensors ve…
-
### Your current environment
vllm version: 0.6.3.post1
### Model Input Dumps
_No response_
### 🐛 Describe the bug
I see on the official site of gemma: https://huggingface.co/google/gemma-2b, cont…
-
Env: torch2.4 cuda 12.4 unsloth main
below is the code errored
```
from unsloth import FastLanguageModel
import torch
model_id="unsloth/gemma-2-2b-it-bnb-4bit"
model, tokenizer = FastLanguageM…
-
Hello. Please add support for Google's open source Gemma AI, which was launched this week.
It has 2 models, 2b and 7b.
Both are great and 2b version can be easily run in both mobile and desktop with…
-
Running lightrag_ollama_demo.py with gemma:2b orlightrag_openai_compatible_demo.py with qwen2.5-3b-instruct wil cast a lot of this:
```
Exception ignored in:
Traceback (most recent call last):
…
-
Hello, I have tried your method on gemma-7b model. I found that this method is work on gsm-8k dataset, but this fails on wikitext-2 dataset. This is my training log:
```
[WARNING|logging.py:329] 2…
-
### What happened?
Model: https://huggingface.co/bartowski/gemma-2-27b-it-GGUF
AMD GPU: RX 7600 XT + RX 7600 (full offload)
With IQ3_M I get about 10 t/s when IQ4_XS is nearly 15 t/s.
I thought …
-
# Fix for gemma-2-9b - run with blfloat16
![image](https://github.com/ObrienlabsDev/machine-learning/assets/24765473/4e149bf2-e84e-48a8-b3bc-1939d1543f66)
https://huggingface.co/google/gemma…