facebookresearch / LAMA

LAnguage Model Analysis
Other
1.36k stars 184 forks source link

The bert-base-cased downloaded from s3 link has something wrong #49

Open CaptXiong opened 2 years ago

CaptXiong commented 2 years ago

The tokenizer.vocab_size is 30522, however it should be 28996 which is just the same with config.vocab_size. I suggust that we can just download the model from hugging face.