ai-forever / ner-bert

BERT-NER (nert-bert) with google bert https://github.com/google-research.
MIT License
407 stars 97 forks source link

should we calculate F1-score with micro-average or macro-average? #26

Open Junpliu opened 4 years ago

Junpliu commented 4 years ago

In the jupyter notebook "conll2003 BERTBiLSTMCRF" in the "examples" folder, the result report is as follow:

image

I notice you put macro-avg "0.9221" in the "README.md" file, but it seems like that the code at "https://paperswithcode.com/sota/named-entity-recognition-ner-on-conll-2003" adopt the micro-avg value as the final F1-value.

I would appreciate it very much if you can tell me why, thanks.