-
There's some work being done to implement Infini-attention from https://arxiv.org/pdf/2404.07143
In a nutshell it allows for essentially an unlimited context length without incurring the quadratic …
-
In the [notebook](https://colab.research.google.com/drive/1fxDWAfPIbC-bHwDSVj5SBmEJ6KG3bUu5?usp=sharing#scrollTo=LjY75GoYUCB8) where you mentioned about how absence of `` token affects the training lo…
AvisP updated
3 months ago
-
Hi, I am trying to use `TestsetGenerator` to produce a synthetic dataset paired with `LlamaIndex` and 'Ollama', it successfully completes the embedding process, but before startin the generation proce…
-
### Feature request
add option to stream output from pipeline
### Motivation
using `tokenizer.apply_chat_template` then other stuff then `model.generate` is pretty repetitive and I think it's time …
-
I made a colab( https://colab.research.google.com/drive/1Z3NdoT0WS8KXnSUS3_xxT39NBZD6eGcN?usp=sharing ) to test and I ran into some issue. GemmaModel.forward() got an unexpected keyword argument 'ca…
-
Hi, I'm using the lighteval to test several benchmarks, but I met issues when testing the following 2 benchmarks:
1. lighteval|wikitext|0|0
2. helm|wikitext:103|0|0
When test wikitext, I got er…
-
### System Info
`peft=0.12.0`
`transformers =4.44.0`
### Who can help?
_No response_
### Information
- [ ] The official example scripts
- [x] My own modified scripts
### Tasks
-…
-
Hi there, I encountered a strange bug after trying to load the gemma-2b model using kerasnlp.
My finetuning code is the following:
` def fine_tune(self, X, y):
data = generate_train…
-
-
how to finetune with gemma model?