wietsedv / bertje

BERTje is a Dutch pre-trained BERT model developed at the University of Groningen. (EMNLP Findings 2020) "What’s so special about BERT’s layers? A closer look at the NLP pipeline in monolingual and multilingual models"
https://aclanthology.org/2020.findings-emnlp.389/
Apache License 2.0
133 stars 10 forks source link

NER Labels #27

Closed rensdebee closed 2 years ago

rensdebee commented 2 years ago

When trying to parse a dutch sentence using transformers NER the only output labels that exist are label_0 and label_1, is this correct and what do these labels mean?

wietsedv commented 2 years ago

You are likely using the pre-trained model (GroNLP/bert-base-dutch-cased). This model is not fine-tuned for a task like NER. You'll want to fine-tune the model on some labelled data or use a model that is already fine-tuned for NER, like wietsedv/bert-base-dutch-cased-finetuned-conll2002-ner. You can see in the config that there are labels.