-
### Describe the bug
1) your ui initial start is like this (perfect):
2) whenever you open additional inputs and then closed then height end like this:
in terms of UX is no functional.
…
-
Since the latest models, such as Llama 3 and Gemma, adopt extremely large vocabularies (128-256K), the size of logits can become very large, consuming a large proportion of VRAM. For example, the foll…
-
Hello Authors,
Thank you for your incredible work and the comprehensive experiments presented in the paper.
I have a question regarding the implementation of attacks. Specifically, some attacks,…
-
Please add Qwen2 support
```
EETQ_CAUSAL_LM_MODEL_MAP = {
"llama": LlamaEETQForCausalLM,
"baichuan": BaichuanEETQForCausalLM,
"gemma": GemmaEETQForCausalLM
}
```
-
I found that the current repository configuration is not compatible with Gemma2. The reason might be that transformers and vllm are not fully compatible with Gemma2. Could you share the package config…
-
Compile the database of learning materials that contains longer and more reasonable materials.
-
### Description of the bug:
I ran the Gemema-7B model based on the code in the example, and found that the model's answers were rather poor and didn't seem to understand my question at all. Is this …
-
hello. I am getting an error when running the sample below.
The request file does not exist in the original source,
I copied and used the preprocessor_config.json file in the same model family.
…
-
Edited by PP to add that this issue was noted during splitting an experiment, which turned out to be misannotated so it's not a problem, but in general, when we have FactorValues with multiple stateme…
-
URLs are created relative to settings in `Gemma.properties`, but those are redundant since it is also part of the servlet context configuration. It would also be more robust and work consistently in d…