richardbaihe / paperreading

NLP papers
MIT License
2 stars 0 forks source link

Arxiv 2020| GMAT: Global Memory Augmentation for Transformers #35

Closed richardbaihe closed 4 years ago

richardbaihe commented 4 years ago

This paper proposes the chunked self-attention with a global memory module.

image