ymcui / Chinese-BERT-wwm

Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型)
https://ieeexplore.ieee.org/document/9599397
Apache License 2.0
9.56k stars 1.38k forks source link

有点晕:『开源版本不包含MLM任务的权重』,这句话是什么意思? #226

Closed liangcaihua closed 1 year ago

liangcaihua commented 1 year ago

文中提到了一句:

注意:开源版本不包含MLM任务的权重;如需做MLM任务,请使用额外数据进行二次预训练(和其他下游任务一样)

我理解BERT的预训练模型会使用MLM和NSP两种训练模式,并且主要使用了MLM训练,那么不包含MLM任务的权重是啥意思?

stale[bot] commented 1 year ago

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

stale[bot] commented 1 year ago

Closing the issue, since no updates observed. Feel free to re-open if you need any further assistance.

chansonZ commented 3 months ago

同问