clamsproject / aapb-evaluations

Collection of evaluation codebases
Apache License 2.0
0 stars 1 forks source link

More improvements to NEL evaluation #30

Open wricketts opened 10 months ago

wricketts commented 10 months ago

Because

Latest metrics in the 20230824 evaluation are pretty low. This could be due to several reasons:

It could be insightful to add a more fine-grained evaluation for each annotation property. Specifically, by computing precision, recall, and F1 for (some options)--

If metrics are particularly low for one of these compared to others, it might show where the app could be improved.

Done when

More fine-grained evaluation is implemented (or we decide it's not necessary).

Additional context

No response

keighrim commented 6 months ago

TIL about this; https://www.semantic-web-journal.net/system/files/swj1671.pdf Maybe we need to take a closer look at the library and consider using standardized metrics included.