flrngel / understanding-ai

personal repository
36 stars 6 forks source link

Bi-Directional Block Self-Attention for fast and memory-efficient sequence modeling #2

Open flrngel opened 6 years ago

flrngel commented 6 years ago

Paper at ICLR2018 https://openreview.net/forum?id=H1cWzoxA- aka BiBloSAN (Bi-BloSAN)

Abstract

1. Introduction

2. Background

2.2 Vanilla Attention and Multi-dimensional Attention

Bi-BloSAN

image

4. Experiments

Terminology