ymcui / Chinese-BERT-wwm

Pre-Training with Whole Word Masking for Chinese BERT(中文BERT-wwm系列模型)
https://ieeexplore.ieee.org/document/9599397
Apache License 2.0
9.56k stars 1.38k forks source link

RoBERTa-wwm-ext-large ft的时候loss飞了 #232

Closed fword closed 1 year ago

fword commented 1 year ago

lr:2e-5 batch_size:16

8966A67C-BCAA-4CA6-BF2C-FF4CE3182912

stale[bot] commented 1 year ago

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

stale[bot] commented 1 year ago

Closing the issue, since no updates observed. Feel free to re-open if you need any further assistance.