-
### Have I written custom code (as opposed to using a stock example script provided in MediaPipe)
None
### OS Platform and Distribution
IOS
### MediaPipe Tasks SDK version
_No response_
### Task…
-
Currently, `evaluation.yaml` exists under the `configs/` directory. To start, we wanted to just showcase this recipes as an example, but it is a core part of the finetuning process and therefore shou…
-
FAILED plugins/validation_tests/test_object_creation.py::test_all_suts_can_evaluate[gemma-9b-it-hf] - modelgauge.secret_values.MissingSecretValues: Missing the following secrets:
scope='hugging_face' …
-
Please add Qwen2 support
```
EETQ_CAUSAL_LM_MODEL_MAP = {
"llama": LlamaEETQForCausalLM,
"baichuan": BaichuanEETQForCausalLM,
"gemma": GemmaEETQForCausalLM
}
```
-
### Describe the bug
1) your ui initial start is like this (perfect):
2) whenever you open additional inputs and then closed then height end like this:
in terms of UX is no functional.
…
-
I tried to use ctranslate2 as the inference framework to do model inference, but failed with error as below:
"axis 2 has dimension 8192 but expected 7680"
What I've done:
1. First I must con…
-
### Description of the bug:
I ran the Gemema-7B model based on the code in the example, and found that the model's answers were rather poor and didn't seem to understand my question at all. Is this …
-
I found that the current repository configuration is not compatible with Gemma2. The reason might be that transformers and vllm are not fully compatible with Gemma2. Could you share the package config…
-
修改SGLang 框架对 BAAI/bge-reranker-v2-gemma 模型适配,并进行推理加速
## 测试配置
- **硬件**: NVIDIA A10 GPU
- **序列长度**: 512
## 性能结果
下表对比了不同批量大小下,原始推理与 SGLang 加速推理的时间表现:
此次推理过程中使用的 query 和 document 数据均为随机生成,长度和为…
-
Since the latest models, such as Llama 3 and Gemma, adopt extremely large vocabularies (128-256K), the size of logits can become very large, consuming a large proportion of VRAM. For example, the foll…