Closed bckim92 closed 5 years ago
Hi, Actually, we have tried adding the absolute and relative positional embeddings in our model. But we found those embeddings doesn’t help (accuracy doesn’t improve). So we just deleted them for simplicity.
Okay, thanks for your reply!
Hi, thanks for your great work.
I get to know that there is an architecture difference between SummaRuNNer and extractor of this work. In SummaRuNNer, they add absolute/relative positional embeddings to the classification layer, but in your work, you didn't add those embeddings (related code). Is there any special reason for this architecture choice (e.g. faster training, performance degradation)? If so, can you tell it to me?
Thanks!