-
Hello Authors,
Thank you for your incredible work and the comprehensive experiments presented in the paper.
I have a question regarding the implementation of attacks. Specifically, some attacks,…
-
Hello,
Beam later version is V2 and they did drastic changes to their SDK and client that makes most of the training (fine-tuning) and inference code useless. There is no "beam run" and so on...
…
-
Linear Regression has been implemented on single core.
Exact same output upto 6 decimal places :D .
To test
```sh
./build/faster_lmm_d --geno=data/gemma/BXD_geno.txt.gz --pheno=data/gemma/BXD.ph…
-
Compile the database of learning materials that contains longer and more reasonable materials.
-
### Description of the bug:
I ran the Gemema-7B model based on the code in the example, and found that the model's answers were rather poor and didn't seem to understand my question at all. Is this …
-
Edited by PP to add that this issue was noted during splitting an experiment, which turned out to be misannotated so it's not a problem, but in general, when we have FactorValues with multiple stateme…
-
I tried to reproduce your gemma2B reward model training again and found that the reward model architecture fine tuned with internlm2 had an output header of 1. I downloaded your GRM-Gemma-2B-Sftrug re…
-
URLs are created relative to settings in `Gemma.properties`, but those are redundant since it is also part of the servlet context configuration. It would also be more robust and work consistently in d…
-
hello. I am getting an error when running the sample below.
The request file does not exist in the original source,
I copied and used the preprocessor_config.json file in the same model family.
…
-
Please add Qwen2 support
```
EETQ_CAUSAL_LM_MODEL_MAP = {
"llama": LlamaEETQForCausalLM,
"baichuan": BaichuanEETQForCausalLM,
"gemma": GemmaEETQForCausalLM
}
```