Open sbmaruf opened 5 years ago
Yes, you are right, but under the existing experimental conditions, I can‘t improve the results to about 92.4%. Maybe some tricks need to be used, or some parameters need to be adjusted.
Hi @kyzhouhzau. Any follow-up in this? Have you find the workaround to reproduce the original result of the BERT paper.
@sbmaruf Hi,I want to aks a question.Do this model use the POS in NER?
Excuse me. Where does the label_test.txt come from? Man made or machine generated?
@zwd13122889 See here, https://github.com/kyzhouhzau/BERT-NER/blob/master/BERT_NER.py#L190-L195 It only reads the first token and the last token. In ConLL dataset first token is the text and last token is the label of NER. So POS tag is used.
@sbmaruf OK.How long will it take me to finish this script with a gpu?
It depends on what GPU you are using. I forgot but it should not take more than 2-3 hours in GTX 1080ti/2080ti.
OK. I run my own data. But i have some problem show in the picture:
the left is author's data ,the right is mine
Hi, does anyone know what 98.07% on the first line mean?
Hi, it's accuracy. For multi-class classification accuracy
is not a good measurement.
Oh I see! I was fooled by the format. Thanks a lot!
Your final result seems,
Result description: As Google's paper says a 0.2% error is reasonable(reported 92.4%).
How can this result is comparable to google's result. google's result was 92.4 for BERT base and 92.8 for BERT large. This result is 89.45.