flairNLP / flair

A very simple framework for state-of-the-art Natural Language Processing (NLP)
https://flairnlp.github.io/flair/
Other
13.9k stars 2.1k forks source link

Parameters for pretrained datasets #1943

Closed erlebach closed 3 years ago

erlebach commented 3 years ago

HI,

Where can I find the parameters you used to generate the various pre-trained datasets? I tried to duplicate the POS tagging for the OntoNotes dataset, but only got 60% testing accuracy after 100 epochs. This leads me to believe that I am not using a correct set of hyperparameters. If you could provide this information, I would greatly appreciate it. Thank you. Gordon.

stefan-it commented 3 years ago

Hi @erlebach ,

could you specify which model you would like to replicate? E.g. OntoNotes results are reported for NER, PoS tagging model is available on Wall Street Journal dataset. The parameters should be documented here:

https://github.com/flairNLP/flair/blob/master/resources/docs/EXPERIMENTS.md

Would be great if you could also provide a code snippet that you're using for training; I can have a look at it then :hugs:

cpuyyp commented 3 years ago

Hi @stefan-it ,

I'd like to know the parameters and replicate the fine-grained PoS tagging on OntoNotes dataset, which is reported with a 98.19% accuracy. I found it under the English-models section in the list of pre-trained models. Link here: https://github.com/flairNLP/flair/blob/master/resources/docs/TUTORIAL_2_TAGGING.md#english-models

Really appreciate your help!

alanakbik commented 3 years ago

The English Ontonotes POS model shipped in Flair was trained with these embeddings:

        embeddings = StackedEmbeddings(
            [
                FlairEmbeddings("news-forward", with_whitespace=False),
                FlairEmbeddings("news-backward", with_whitespace=False),
            ]
        )

The rest is the same as the other scripts.

stale[bot] commented 3 years ago

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.