ymcui / Chinese-BERT-wwm

Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型)
https://ieeexplore.ieee.org/document/9599397
Apache License 2.0
9.66k stars 1.39k forks source link

model sizing of ”BERT-wwm-ext, Chinese” and ”BERT-wwm, Chinese” #203

Closed elulue closed 3 years ago

ymcui commented 3 years ago

They are all base-level PLMs. ext means the PLM is trained with extended data. Nothing related to the model size.

elulue commented 2 years ago

Thanks a lot.