-
Hello,
I am very interested in your research and am currently trying to run some experiments based on it. However, I encountered an issue while running the program from the HuggingFace_EncDec directo…
-
Hi, thank you for sharing such impressive work. I'm new for this topic. I met a problem when I try to run the code. I wrote the test code as you guided, showing as follows:
from transformers import…
-
@patrickvonplaten
I have been trying to achieve a bleu score of 31.7 (as reported in the blog and paper for WMT en->de evaluation) using hugging-face model **google/bert2bert_L-24_wmt_en_de** but I…
-
Hi,
I would like to know whether TSDAE procedure is advisable for token classification task? or is it better to go with MLM?
Can the TSDAE training code can be also used with any transformer (en…
-
I have trained a EncoderDecoderModel from huggging face to do english-German translation task. I tried to overfit a small dataset (100 parallel sentences), and use `model.generate()` then `tokenizer.d…
-
https://arxiv.org/pdf/1907.12461.pdf
-
Hey guys. I get no benefit from batching (no speedup whatsoever) Sentence-Transformer.
I would love your opinion on the following situation :
I run inferences on **'bert-base-nli-mean-tokens'** mo…
-
## Environment info
- `adapter-transformers` version: 3.2.1
- Platform: Linux-6.2.0-27-generic-x86_64-with-glibc2.37
- Python version: 3.10.9
- PyTorch version (GPU?): 1.13.1 (GPU)
## De…
-
Hi,
Thanks for providing and presenting this nice work.
As mentioned in your paper, your attention pattern for modeling long sequences can be plugged into any pretrained transformer model.
I wond…
-
# 🚀 Feature request
This "Good second issue" should revisit some of the problems we were having with FP16 for `T5ForConditionalGeneration`: https://github.com/huggingface/transformers/issues/4586 a…