Closed MichaelRoeder closed 10 years ago
Make sure that the licence is not violated when the data is transformed to the annotation backend. Keep in mind that we only want open data and open source software.
similarly to #16, what does open data mean?
TAC KBP 2014 scorer: https://github.com/wikilinks/neleval
I propose to use the scorer as it is in the official repository and formatting both GS and system outputs in order to fit the expected inputs of the scorer.
Few aspects to consider: an entity is defined as an ordered list of the following features: doc_id,startOffset, endOffset,uri,salience,type
For the majority of the systems supported in GERBIL (and in NERD) doc_id, start and end offset, uri are available. Differently for the type. For instance in Babelfy we should retrieve it from a Wikipedia page or so (am I mistaken?). Similarly for the salience score.
Please refine what you mean by experiment types, matchings and evaluation measures and open new and separate issues for all of them
see #48 #49. @giusepperizzo Are there more experiment types we have to cover?
Just to clarify:
Right, but the world is complex.
rdf:type
to an entity, right? Do you say a word with respect to which taxonomy (e.g. schema.org, YAGO, DBpedia-OWL, etc.) OR alignments between taxonomies (e.g. NERD ontology)?better we move the definition to the paper and the discussion to the mailing list.
Add experiment types from http://nlp.cs.rpi.edu/kbp/2014/KBP2014EL_V0.2.pdf