-
### Is there an existing issue for this?
- [X] I have searched the existing issues and did not find a match.
### Who can help?
_No response_
### What are you working on?
As part of a dissertation…
-
### System Info
- transformers version: 4.36.0
- Platform: Linux-4.19.0-22-amd64-x86_64-with-glibc2.31
- Python version: 3.10.13
- Huggingface_hub version: 0.19.4
- Safetensors version: 0.4.0
- …
-
Create a directory with name trained_models or something similar where all trained, fine-tuned models used in the pipeline are stored. So that each time you need them, you can just directly load them …
-
**Describe the bug**
Cannot export the model.
**To Reproduce**
```
import keras
from keras_nlp.models import XLMRobertaPreprocessor, XLMRobertaBackbone
import tensorflow as tf
preprocessor …
-
## モデルの基本情報
**name**: jina-embeddings-v3
**type**: XLMRoBERTa (+ LoRA Adapter)
**size**: 559M (LoRA Adapterを加えると572M)
**lang**: multilingual
## モデル詳細
https://arxiv.org/abs/2409.10173
ht…
-
In `from_pretrained()` in `model.py`, if the `config` is neither `BertConfig`, `RobertaConfig` or `DistilBertConfig`, the `tensor` won't be initialized.
I peeped the codebase and found that `tensor…
-
我在https://huggingface.co/models?search=roberta_large
里面搜素只有这几个,不知道到底是哪个?
ahotrod/roberta_large_squad2 Has a model card
ji-xin/roberta_large-SST2-two_stage
ji-xin/roberta_large-MRPC-two_stage
…
-
When I am training the XLM-Roberta based QE system, I pre-downloaded the pre-trained XLM-Roberta model from huggingface's library and modified the field `system.model.encoder.model_name` in `xlmrobert…
-
Hello,
I am trying to use the XLMRoberta model instead of BERT and I made the following changes to the `bert_pretrained.py`:
```
from transformers import XLMRobertaTokenizer
from transformers …
-
Hi, thanks for the great example on training RoBERTa with long attention.
Followed this example: https://github.com/allenai/longformer/blob/master/scripts/convert_model_to_long.ipynb
Was able to s…