-
## CUDA Out of Memory error but CUDA memory is almost empty
I am currently training a lightweight model on very large amount of textual data (about 70GiB of text).
For that I am using a machine on…
-
Hi, @patrickvonplaten, in the Leveraging Pre-trained Language Model Checkpoints for Encoder-Decoder Models blog post. https://huggingface.co/blog/warm-starting-encoder-decoder
Some of the links to …
-
We are trying to do text summarization using BERT2BERT model. But facing this error mentioned below:
The above exception was the direct cause of the following exception:
KeyError …
-
Hey, I want to load the cnn-dailymail dataset for fine-tune.
I write the code like this
from datasets import load_dataset
test_dataset = load_dataset(“cnn_dailymail”, “3.0.0”, split=“train”)
A…
-
# 🐛 Bug
Hello, thank you for the recent [PR](https://github.com/huggingface/transformers/pull/4436) with fp16 fixes. It seems to work well with short inputs, but once the model is fed with some mor…
-
## Environment info
- `transformers` version: 4.2.1
- Platform: Ubuntu 20.04.1 LTS
- Python version: 3.8.5
- PyTorch version: 1.7.1
- Using GPU in script?: Yes
- Using distributed or paralle…
-
Hello,
I am actually creating this for posterity because it took me a day to figure it out and if anybody else has this issue, hopefully this helps.
I am running a Bert2Bert EncoderDecoderModel …
-
How to train a custom seq2seq model with `BertModel`,
I would like to use some Chinese pretrained model base on `BertModel`
so I've tried using `Encoder-Decoder Model`, but it seems the`Encoder-…
-
Hello everyone, I need help with the training of the encoder-decoder model. I need to fine-tune a bert2bert for Turkish content summarization. I am using this sample notebook reference: https://githu…
-
Hello ! I am trying to launch an `accelerate launch` but I am encounting those error:
```Exception: Installed CUDA version 11.7 does not match the version torch was compiled with 11.6, unable to co…