ELECTRA: Pre-training Text Encoders as Discriminators Rather Than Generators
2.33k
stars
352
forks
source link
The exact English pretraining data and Chinese pretraining data that are exact same to the BERT paper's pretraining data. #118
Open
guotong1988 opened 3 years ago
Any one know where to get them? Thank you and thank you.