wangqiangneu / MT-PaperReading

Record my paper reading about Machine Translation and other related works.
36 stars 2 forks source link

20-ICLR(reviewing)-ELECTRA: PRE-TRAINING TEXT ENCODERS AS DISCRIMINATORS RATHER THAN GENERATORS #25

Open wangqiangneu opened 5 years ago

wangqiangneu commented 5 years ago

简介

discriminator代替BERT中的MLM generator. 具体方法是,用一个small MLM generator把输入x改造成$\hat{x}$ (随机mask 15%,用这个generator生成masked token,有可能恢复原样,有可能跟原始输入不同)。用discriminator去二分类$\hat{x}$中的所有token,是不是原始的输入。训练时generatordiscriminator联合学习(word embedding是shared,其他参数不共享)。然后finetuning时,只使用discriminator,不用generator了。

论文信息

总结