-
The cookbook aims to provide a comprehensive guide for researchers and practitioners interested in fine-tuning the Gemma model from Google on a mental health assistant dataset.
Key components of th…
-
It's effectively being used broader than the VNG/GEMMA & Zaakgericht Werken context - it really is aimed at JSON-based OpenAPI 3 driven services.
Proposal: `oas3-client` (boring but carries the wei…
-
### 🚀 The feature, motivation and pitch
Gemma-2 and new Ministral models use alternating sliding window and full attention layers to reduce the size of the KV cache.
The KV cache is a huge inferen…
-
David+Whitney
Alexis+Anacona
Tracey+Sazare
Atlas+Siluca
Damar+Aishela
Worapoj+Lalrinkimi
Hector+Harmonica
Takeshi+Kasumi
Wayne+Lazuli
Arjen+Ketifa
Kanatbek+Lamia
Kazuhiro+Shinobu
Tugstsogt…
-
- [x] MiniCPM-Llama3-V-2_5
- [x] Florence 2
- [x] Phi-3-vision
- [x] Bunny
- [x] Dolphi-vision-72b
- [x] Llava Next
- [x] Qwen2-VL
- [x] Pixtral
- [x] Llama-3.2
- [x] Llava Interleave
- [x] …
-
Hi.
I've been fine-tuning Gemma-2-2B-it on Google Colab, saved the fine-tuned model to Huggingface.
When I load the model from Huggingface hub, I keep getting inference errors.
`from unsloth impo…
-
**Qwen2**
warning: not compiled with GPU offload support, --n-gpu-layers option will be ignored
warning: see main README.md for information on enabling GPU BLAS support
Log start
main: build = 2…
-
Hi @danielhanchen
I am trying to fine-tune gemma2-2b for my task following the guidelines of the continued finetuning in unsloth. Howver, I am facing OOM while doing so. My intent is to train gemm…
-
### Name and Version
version: 4293 (3d98b4cb)
built with MSVC 19.42.34435.0 for x64
### Operating systems
Windows
### Which llama.cpp modules do you know to be affected?
Python/Bash scripts
###…
-
Nice project. Open source models are usually good at one task, such as coding, writing, etc. And it would be interesting if we could set a parameter in the .env to specify which model that agent shoul…