-
I wanted to check the feasibility of implementing roberta-large-mnli model through ggml. Is there anything that could be a potential hurdle ?
-
Why did you use Roberta and not use BERT or ELMO instead?
-
Hello,
First, congratulations for your work.
Second, from what I have discovered so far, you only allow Bert like training and not Roberta training.
Even if the NSP is set to false, still your …
-
Hey
Does this work with FacebookAI/xlm-roberta-large-finetuned-conll03-english too? And where can I find base_vocabulary.json?
-
Hi,
I use transformers==4.28.1 and torch==2.1.0
I run the following command:
```shell
TASK=SNLI K=512 SEED=42 BS=64 LR=1e-4 EPS=1e-3 STEP=50000 MODEL=roberta-large EXTRA_TAG=lora bash mezo.sh --…
-
### 🐛 Describe the bug
Category | Model | Accuracy
-- | -- | --
torchbench_amp_bf16_training | basic_gnn_edgecnn | eager_two_runs_differ
torchbench_amp_bf16_training | …
-
To whom this may concern,
Thank you for the work to make these examples available. I was working through the Sentiment Analysis with Pinecone Google Colab notebook and found that the model_id for t…
-
RoBERTa는 SOTA모델로 중요한 모델 중 하나입니다.
더 높은 f1 score를 뽑아냅시다.
-
1. I want to do incremental pre-training on the existing RoBERTa. Which RoBERTa model should I use? Download directly from Hugging Face? Do I need to script it into UER format after downloading it? Is…
-
Could you add xlm roberta?
see https://huggingface.co/transformers/pretrained_models.html