Closed richardbaihe closed 4 years ago
https://github.com/allenai/longformer
This paper proposes local+global attention for the transformer.
achieves SOTA on WikiHop and TriviaQA after pretraining Roberta with Longformer and finetuned with Longformer.
https://github.com/allenai/longformer
This paper proposes local+global attention for the transformer.
achieves SOTA on WikiHop and TriviaQA after pretraining Roberta with Longformer and finetuned with Longformer.