cl-tohoku / bert-japanese

BERT models for Japanese text.
Apache License 2.0
514 stars 55 forks source link

Getting some weights not used warning #19

Open wailoktam opened 4 years ago

wailoktam commented 4 years ago

Hi, I get the following warning when loading a checkpoint trained using your pretrained model. Is there anything wrong with it?

Some weights of the model checkpoint at /content/drive/My Drive/pretrainedBertJa/bert-base-japanese-whole-word-masking were not used when initializing BertForSequenceClassification: ['cls.predictions.bias', 'cls.predictions.transform.dense.weight', 'cls.predictions.transform.dense.bias', 'cls.predictions.transform.LayerNorm.weight', 'cls.predictions.transform.LayerNorm.bias', 'cls.predictions.decoder.weight', 'cls.seq_relationship.weight', 'cls.seq_relationship.bias']

singletongue commented 3 years ago

This IS an expected behavior since you are initializing a BertForSequenceClassification model from our BertForPretraining model. See the warning message in the first bullet point.