Closed puraminy closed 2 years ago
Hi,
Could you check your transformers version and datasets version and make sure they are the same with we provided?
Thanks
I run it on conda virtual env. and yes they are the same
transformers==4.9.2
datasets==1.14.0
Ok, then it seems like a problem caused by RAM?
Hi,
The separate
feature is deprecated and not used in our experiment, so we did not test it. In our latest commit e337e3b0b492772ba10f35bf64112c10d16d203c
, we attempted to fix this bug. In short, the model_max_length
flag of T5 is set as a very large integer, which causes RAM overflow. Could you check the latest code? Thanks!
However, you may face GPU memory overflow since the input length is effectively doubled when using separate
.
I manually changed the utils/dataset.py according to your commit. However self.tokenizer.input_max_length
isn't recognized. I fixed it to 100
and now the problem was resolved.
By the way, what is the difference of separate
and concatenate
setting?
Hi, thanks for pointing out, we will fix that sooner.
For the meanings of separate
and concatenate
version:
separate
: Get the prefix weight from query
, structured knowledge
separately but has been deprecated by us since it didn't improve the result in our early version.
concat
: concat the query
, structured knowledge
and context
to formulate them as one sentence input and get its prefix weight. It is simple and effective thus we adopt it as defaulted.
Contact us if you have some further findings!
in fetaqa config file if I change
concatenate
toseparate
and run prefix tuning the following error occurs