-
Thanks for the great work. I'm trying to reproduce the results and facing following errors:
1. Can I use lm-evaluation-harness script instead of yours to evaluate the results? When I used lm-harness …
-
While generating any text with a specified value of max_length, the generated text keeps repeating several times until the output spans the value of max_length. An example of the above is using the fo…
-
### Model ID
google/gemma-2-2b
### Model type
Decoder model (e.g., GPT)
### Model languages
- [x] Danish
- [x] Swedish
- [x] Norwegian (Bokmål or Nynorsk)
- [x] Icelandic
- [x] Faroese
- [x] Germ…
-
### System Info
Transformers v4.45.0.dev0
### Who can help?
@Rocketknight1
### Reproduction
The code snippet from [here](https://huggingface.co/docs/transformers/main/en/chat_templatin…
-
Thank a lot for your great work!
I deployed gemma-2b locally. I would like to understand how to have multiple rounds of dialog effectively.
I searched the internet and found that I could type in p…
-
I have been trying llm_inference on android (https://github.com/google-ai-edge/mediapipe-samples/tree/main/examples/llm_inference/android). While the model performs perfectly in the emulator, it start…
-
This error occurs with gemini:
```
Error executing vim.schedule lua callback:
called `Result::unwrap()` on an `Err` value:
RequestError(Status(401, Response[status: 401, status_text: Unauthoriz…
-
Seems like a couple leading options for running LLM inference locally are ollama and llama.cpp. post & discussion [here](https://www.reddit.com/r/LocalLLaMA/comments/1d05x6v/llamacpp_runs_18_times_fas…
-
If you are submitting a bug report, please fill in the following details and use the tag [bug].
**Describe the bug**
Gemma-2-{size} is not loadable using from_pretrained. I checked OFFICIAL_MODEL_…
-
Implement a correctness aware loss that is like Yuchi's but uses weighting according to accuracy
```math
L(\vec{c}, \vec{t}) = \frac{1}{n}\sum_{i = 1}^n f(t_i)(c_i - t_i)^2
```
where
```math
f(…