-
I'm running on WSL2/Ubuntu on Win11. Deliberately using CPU mode as my GPU is too weak. Using Python 3.10.12.
Here is the output when trying to run sampling.py:
```
~/gemma$ python3 examples/sa…
-
### System Info
```shell
Optimum version: d87efb2
Transformers version: d479665
ONNXRuntime version: 1.17.1
ONNX version: 1.15.0
```
### Who can help?
@michaelbenayoun @echarlaix
### Informat…
-
## Bug report
**Describe the bug**
LLM Engine failed in ValidatedGraphConfig Initialization step.
### Steps to reproduce
Steps to reproduce the behavior:
1. Download gemma-2b-it-gpu-int8.…
-
Hello, please upgrade Google Gemma models to version 1.1 and include it in Android prebuild app.
Links to models:
https://huggingface.co/google/gemma-1.1-2b-it
https://huggingface.co/google/gemma-1.1…
-
# Experiments
Idea: Repeat most of the unlearning experiments (continuous, batch, sequential) with harmfulness and evaluate. Based on the results decide the best hyperparameters for unlearning fren…
-
I am using "google/gemma-2b-it" model from HuggingFace. I realized there are 99 unused tokens (\ ,\,\...) in first 106 token ids. Does anyone know their purpose? Just wondering.
-
# 模型参数支持专区
大家好,PaddleNLP 团队在这里为大家整理了各个模型参数的详细信息,方便大家使用。
## 模型参数
### Base Models
| Model | 0.5B | 1~2B | 3~4B | 6~7B | 13~14B | 30~32B | 50~60B | 65~72B | 110B | >110B |
|:---------:|:--…
-
**Describe the bug**
When attempting to shard a `gemma_2b_en` model across two (consumer-grade) GPUs, I get:
```
ValueError: One of device_put args was given the sharding of NamedSharding(mesh=…
-
### Feature request
I see [llama](https://github.com/huggingface/transformers/blob/main/src/transformers/models/llama/modeling_llama.py#L829-L835) will remove tuple past key values in 4.43.
### Moti…
-
Hi 👋 ,
It would be really great if you could add support for the Gemma model series (i.e. 2B and 7B variants, particularly the 7B is what I would like most), since I see that it is currently not su…