-
Hi everyone.
I see the issue when run unit tests on NVidia A100 (GPU). Here is the [link](https://github.com/NVIDIA/JAX-Toolbox/actions/runs/9099672951/job/25013673011?pr=590#step:7:593) for more d…
-
I'll beautify this once I get hold of Azure storage.
I have attached [gemma_7b.mlir](https://storage.googleapis.com/shark_tank/dan/Gemma/gemma_7b.mlir) along with [gemma weights](https://storage.go…
-
Hello. Please add support for Google's open source Gemma AI, which was launched this week.
It has 2 models, 2b and 7b.
Both are great and 2b version can be easily run in both mobile and desktop with…
-
### What happened?
Model: https://huggingface.co/bartowski/gemma-2-27b-it-GGUF
AMD GPU: RX 7600 XT + RX 7600 (full offload)
With IQ3_M I get about 10 t/s when IQ4_XS is nearly 15 t/s.
I thought …
-
I think this is related to #4 but I open a new issue because it seems something is changed after it.
I'm using likelovewant/ollama-for-amd 0.3.6
I already tried various combination of the binaries i…
-
Authentication in code with token=hf_token doesn't work unless you use subprocess.run("local-gemma", "--token", hf_token, "What is the capital of France")
`model = LocalGemma2ForCausalLM.from_pretr…
-
### Your current environment
```text
PyTorch version: 2.3.0+cu121
Is debug build: False
CUDA used to build PyTorch: 12.1
ROCM used to build PyTorch: N/A
OS: Ubuntu 22.04.3 LTS (x86_64)
GCC ve…
-
Hi,
I'm facing the following issue when trying to chat with Ollama:
```
04/17/2024 01:13:07 PM utils.py 273 : Failed to get max tokens for LLM with name gemma. Defaulting to 4096.
Trac…
-
### Model description
Gemma is the recent open-weights family of models that Google released
### Open source status
- [ ] The model implementation is available
- [X] The model weights are av…
-
### What happened?
I'm using local Ollama with Code Gemma model. This works fine sending chats and code explain etc. However when I select generate for the check in submit it simply crashes and I can…