Closed drxmy closed 2 years ago
Hi,
Your understanding is right and that's indeed an interesting setting! We planned to do such experiments and add that to Table 4 of the paper, but we thought tuning foundation model like T5 doesn't fit in our blueprint which assume the super-large-model shouldn't be changed thus we didn't do so actually.(And of course we may make it up if reviewers require us to do so =)
Hope this information helpful! Thanks
Thank you for the clarification.
When doing the Multi-task prefix-tuning, in my understanding the T5 is not finetuned on any SKG task(or not finetuned at all?). Would it be better finetune it first and then prefix tuning it again? Just curious whether you have done some experiment about this.
Really wonderful work!