-
I encountered an issue while trying to reproduce the results by loading the gpt2_bc_webshop_history.pt model and running the run.py script. The training was initiated with the following parameters:
…
-
I noticed that GPT2Tokenizer is used when evaluating GPT2, which doesn't have a mask_token. Will this impact the evaluation result?
I think I should add a new one manually but I'm unsure which one I…
-
-
Python is doing a lot of heavy lifting and hiding a lot of the complexity. It'd be interesting to compare the Python GPT code with Kaparthy's pure C/CUDA implementaton of the same:
https://github.c…
-
**The bug**
When using `models.LlamaCpp` the selected tokenizer is always gpt2 (This can be seen in the outut when `verbose=True` arg is set). I have pasted the dumped KV metadat keys
```
llama_mod…
-
Trying to run following gpt2 demo with triton-cpu and certain tirton kernel fails with above error:
```python
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
from torch._…
-
-
!pip install transformers datasets
from transformers import GPT2Tokenizer, GPT2LMHeadModel, Trainer, TrainingArguments
from datasets import load_dataset, load_metric
tokenizer = GPT2Tokenizer.from_…
-
### System Info
transformers 4.40.0
python 3.10
### Who can help?
@ArthurZucker
@Narsil
@SunMarc
### Information
- [ ] The official example scripts
- [x] My own modified scripts
…
-
```python
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
@torch.compile(backend="turbine_cpu")
def test_gpt2_demo():
tokenizer = AutoTokenizer.from_pretrained("gp…