-
### Your current environment
```
Kaggle.
```
### 🐛 Describe the bug
```
WARNING: Casting torch.bfloat16 to torch.float16.
WARNING: Gemma 2 uses sliding window attention for every odd la…
-
### ⚠️ Please check that this feature request hasn't been suggested before.
- [X] I searched previous [Ideas in Discussions](https://github.com/OpenAccess-AI-Collective/axolotl/discussions/categories…
-
### Do you need to file an issue?
- [ ] I have searched the existing issues and this bug is not already filed.
- [ ] My model is hosted on OpenAI or Azure. If not, please look at the "model providers…
-
The SimPO is great!
When I evaluated gemma2 on AE2, I did not find the template of gemma2. Could you update the template of gemma2?
-
### What is the issue?
When either steting the OLLAMA_NUM_PARALLEL varaible to value greater than 1 in ollama.service or in docker run command (`sudo docker run -d --gpus=all -v ollama:/root/.ollam…
-
### 🚀 The feature, motivation and pitch
Will it be supported? just saw the new release you uploaded the post1 release.
### Alternatives
.
### Additional context
. For Gemma 2 models
-
### System Info
```
accelerate 0.33.0
peft 0.12.0
Python 3.12.5
macOS: 15.0
MacBook pro M1 Pro 16 gb
```
### Who can help?
### Information
- [ ] The officia…
-
### Your current environment
how to initiate the gemma2-27b with a 4-bit quantization?
### How would you like to use vllm
Could you please explain how to initiate the gemma2-27b with a 4-bit quanti…
-
**Bug Description**
Unable to configure LiteLLM as the provider pointed to a locally running ollama server. It is possible this is user error and I don't have the provider configured correctly, but I…
-
### Description
I've been using gemma-2-9b-it-GGUF in LM studio and it's working a lot better for me than Meta-Llama-3-8B-Instruct-GGUF. I really like it but I'm getting this error when I try to use…