-
Compile the database of learning materials that contains longer and more reasonable materials.
-
I found that the current repository configuration is not compatible with Gemma2. The reason might be that transformers and vllm are not fully compatible with Gemma2. Could you share the package config…
-
Hello,
Beam later version is V2 and they did drastic changes to their SDK and client that makes most of the training (fine-tuning) and inference code useless. There is no "beam run" and so on...
…
-
## 🐛 Bug
Not sure if this is a feature request or bug. I took the [SPMD Gemma ft code from Hugging Face](https://huggingface.co/google/gemma-7b/blob/main/examples/example_fsdp.py) and tried to run …
-
hello. I am getting an error when running the sample below.
The request file does not exist in the original source,
I copied and used the preprocessor_config.json file in the same model family.
…
-
### Description of the bug:
I ran the Gemema-7B model based on the code in the example, and found that the model's answers were rather poor and didn't seem to understand my question at all. Is this …
-
Linear Regression has been implemented on single core.
Exact same output upto 6 decimal places :D .
To test
```sh
./build/faster_lmm_d --geno=data/gemma/BXD_geno.txt.gz --pheno=data/gemma/BXD.ph…
-
Edited by PP to add that this issue was noted during splitting an experiment, which turned out to be misannotated so it's not a problem, but in general, when we have FactorValues with multiple stateme…
-
I tried to reproduce your gemma2B reward model training again and found that the reward model architecture fine tuned with internlm2 had an output header of 1. I downloaded your GRM-Gemma-2B-Sftrug re…
-
URLs are created relative to settings in `Gemma.properties`, but those are redundant since it is also part of the servlet context configuration. It would also be more robust and work consistently in d…