Closed abinpaul1 closed 5 years ago
When you're calling the script, are you providing an existing model, or are you training from a blank model?
If you're training from a blank model, then I think the problem would be you're trying to learn labels that aren't in the model. If you're starting from a pretrained model, then I'm less sure what could be wrong. It could be that it's struggling to update from so few texts, maybe try setting the batch size to 1? I'm not sure though, it should learn something.
Ok. So I increased the number of examples to 70, added all the labels and trained a new model. Now it's varying in between 40-70.
If I train a new model , it wouldn't have all other features like parts of speech tagger an all of the original modal right. Should I use the new model exclusively to tag these particular entities.
If I train a new model , it wouldn't have all other features like parts of speech tagger an all of the original modal right. Should I use the new model exclusively to tag these particular entities.
If you start off with a blank model, then it won't. But you can also use a pre-trained model that already has a "tagger"
and "parser"
, and only add a new "ner"
component.
This thread has been automatically locked since there has not been any recent activity after it was closed. Please open a new issue for related bugs.
I also posted this question in stack overflow. https://stackoverflow.com/questions/56082191/losses-in-ner-training-loop-not-decreasing-in-spacy
I am trying to train a new entity type 'HE INST'--to recognize colleges. That is the only new label. I have a long document as raw text. I ran NER on it and saved the entities to the TRAIN DATA and then added the new entity labels to the TRAIN_DATA( i replaced in places where there was overlap).
The training loop is constant at a loss value(~4000 for all the 15 texts) and (~300) for a single data. Why does this happen, how do I train the model properly. I have around 18 texts with 40 annotated new entities. Even after all iterations, the model still doesn't predict the output correctly.
I haven't changed the script much. Just added en_core_web_lg, the new label and my TRAIN_DATA
I am trying to tag institutes from resume(C.V) data:
This would be one of my text in TRAIN_DATA: (soory for the long text) I have around 18 such texts concatenated to form TRAIN_DATA
TRAIN DATA EXAMPLE.txt
Also if decide to train new custom NER (for resume entities---Institute,Programming Language,Skill) with a blank 'en' model, won't the parser,tagger,vocab of that model be really bad. How can I mitigate this?
Your Environment
The script I use to do the training: