Closed MonliH closed 1 year ago
Hi @MonliH , Thank you for your interest in our work!
Sorry, our current implementation supports only seq2seq tasks (such as summarization) and does not support QA datasets, because they require encoding the question with each of the input chunks, which is slightly different.
Best, Uri
Hi @MonliH ,
Running with bart_base_sled.json
activates the SLED model, which we based our code on, but is an unrelated approach. We recently found that you can get significant improvements on QA tasks if you concatenate the question + document
and feed them as the input. We will soon update the results in the paper as well.
I'm closing this issue, but feel free to re-open if you have any questions!
Best, Uri
Hello, I am running the bart_base_sled code using the contract_nli dataset, with the following arguments:
(I have low eval_steps for debugging purposes). The training seems to work fine, but after evaluation starts I get an error:
Full Log
``` ╭─────────────────────────────── Traceback (most recent call last) ────────────────────────────────╮ │ /home/aaa/unlimiformer/src/run.py:1213 in