allenai / rainbow

Neural models of common sense. 🤖
Apache License 2.0
94 stars 6 forks source link

is the phenomenon that sequential training hurts performance reasonable? #46

Closed xcluo closed 2 years ago

xcluo commented 3 years ago

Phenomenon: from the cost equivalent curve of Figure 2, sequential training uniformly outperform single task (e.g., target task is winogrande), but in my reproducing, the single task baseline always outperforms all sequential training (sequential training update steps vary from 5k to 50k, and interval is 5k), which violates the UINCORN Table 1 conclusion. download Experiment Setting:

Question:

  1. is the phenomenon that sequential training hurts performance reasonable?
  2. is the margin (77.0%-74.51%=2.49%) compared to the UNICORN single task tolerable?
  3. could you please comment some precious suggestions for us to further improve performance?
xcluo commented 3 years ago

@rlebras

nalourie-ai2 commented 2 years ago

Sequential training did not hurt performance in any of our experiments. However, while our tasks covered a broad range of domains, they were all multiple choice.

In the extreme, sequential training for long enough on totally random data seems like it would hurt performance. Based on the Unicorn experiments, I expect that sequential training will not hurt performance in most practical situations--but as an empirical finding, it's always possible there are important scenarios our experiments didn't cover.