-
## 🐛 Bug
Not sure if this is a feature request or bug. I took the [SPMD Gemma ft code from Hugging Face](https://huggingface.co/google/gemma-7b/blob/main/examples/example_fsdp.py) and tried to run …
-
hello. I am getting an error when running the sample below.
The request file does not exist in the original source,
I copied and used the preprocessor_config.json file in the same model family.
…
-
### Description of the bug:
I ran the Gemema-7B model based on the code in the example, and found that the model's answers were rather poor and didn't seem to understand my question at all. Is this …
-
Linear Regression has been implemented on single core.
Exact same output upto 6 decimal places :D .
To test
```sh
./build/faster_lmm_d --geno=data/gemma/BXD_geno.txt.gz --pheno=data/gemma/BXD.ph…
-
Edited by PP to add that this issue was noted during splitting an experiment, which turned out to be misannotated so it's not a problem, but in general, when we have FactorValues with multiple stateme…
-
URLs are created relative to settings in `Gemma.properties`, but those are redundant since it is also part of the servlet context configuration. It would also be more robust and work consistently in d…
-
Please add Qwen2 support
```
EETQ_CAUSAL_LM_MODEL_MAP = {
"llama": LlamaEETQForCausalLM,
"baichuan": BaichuanEETQForCausalLM,
"gemma": GemmaEETQForCausalLM
}
```
-
### What is the issue?
I tried to import finetuned llama-3.2-11b-vision, but I got "Error: unsupported architecture."
In order to make sure my model is not the problem, I downloaded [meta-llama/Ll…
-
I found that the current repository configuration is not compatible with Gemma2. The reason might be that transformers and vllm are not fully compatible with Gemma2. Could you share the package config…
-
Hi Daniel,
Thank you so much for releasing such an awesome VLM fine-tuning notebook to the public!
I was really excited, tried the notebook out and found the following error:
![image](https:/…