Closed avacaondata closed 3 years ago
That example only runs with BERT
, which is why it has been moved to a separate research project.
I tried this script with albert and it worked, which script should I use to train a Roberta model from scratch with Whole word Masking??
Is that intended:
--model_type deberta
?
@alexvaca0
Sorry, that was from the previous launch script, now it is roberta @cronoik
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
Please note that issues that do not follow the contributing guidelines are likely to be ignored.
Environment info
transformers
version: 4.4.dev0Who can help
@patrickvonplaten @LysandreJik @
Information
When I try to use the BPE Tokenizer trained with huggingface/tokenizers with Roberta directly, it works:
However, when I try to use this same tokenizer for training a language model, it fails:
The error message is the following:
Why doesn't it fail when I try to load the tokenizer with RobertaTokenizer.from_pretrained() but it does fail when I try to run run_mlm_wwm.py ? @sgugger @patrickvonplaten @LysandreJik