-
**Describe the bug**
A clear and concise description of what the bug is.
UD 2.10 数据集训练报错
**Code to reproduce the issue**
Provide a reproducible test case that is the bare minimum necessary t…
-
## Environment info
- `transformers` version: latest (4.10.0.dev0)
- Python version: 3.8
- PyTorch version (GPU?): 1.9.0
- Using GPU in script?: no
- Using distributed or parallel set-up in s…
-
Taking text as example, the transforms may need to download meta-data for it's construction. For example, XLMR transforms requires spm tokenizer model and corresponding vocab to create xlmr preset tra…
-
## Environment info
- `transformers` version: 3.5.1
- Platform: Linux-4.4.0-140-generic-x86_64-with-glibc2.10
- Python version: 3.8.8
- PyTorch version (GPU?): 1.6.0 (True)
- Tensorflow versi…
-
Because I have never trained the pre-training model, I have a small question about what the paralleldata input format looks like`TRAIN_FILE=/path/to/train/file`. Do you need a separator between src an…
-
# 🚀 Feature request
Could it be possible to add adapters to the EncoderDecoderModel?
I am fine-tuning an EncoderDecoderModel with two mBERT and I would like to compare it with fine-tuning only the …
-
# Experiments design
Follow discussion [here](https://docs.google.com/document/d/110tlidAcpiNteKnA27tR5KPS_VahNqYKqCeJlu1MWww/edit#heading=h.wmf5tyes1tfk)
## pointers to code and datasets
### …
-
## Environment info
Copy-and-paste the text below in your GitHub issue and FILL OUT the two last points.
- `transformers` version: 4.8.2
- Platform: macOS-10.15.7-x86_64-i386-64bit
- Pytho…
-
Hi,
If I use the XLMR tokenizer to tokenize, I find there are some unexpected results.
```python
json = {
"text": "Anyhow the man comes in.",
"tokens": ["Anyhow", "the", "man", "…
-
- [ ] Use multilingual pre-trained encoder
- [ ] Train complete network on dutch templates
- [ ] Evaluate on english test templates
Resources:
https://github.com/pytorch/fairseq/tree/master/exam…