AkiraTOSEI / ML_papers

ML_paper_summary(in Japanese)
5 stars 1 forks source link

SpanBERT: Improving Pre-training by Representing and Predicting Spans #110

Open AkiraTOSEI opened 3 years ago

AkiraTOSEI commented 3 years ago

TL;DR

While BERT hides one token (like a word) with a mask and fills it in, Span-BERT hides consecutive tokens (like an idiom) and fills them in. In addition, it does not perform Next Sentence Prediction and only makes the user input a single sentence. It outperforms BERT in many tasks. £(foot ball) = LMLM(football) +  LsBo (football)

Why it matters:

Paper URL

https://arxiv.org/abs/1907.10529

Submission Dates(yyyy/mm/dd)

2019/07/24

Authors and institutions

Mandar Joshi, Danqi Chen, Yinhan Liu, Daniel S. Weld, Luke Zettlemoyer, Omer Levy

Methods

Results

Comments