Closed dineshkh closed 3 years ago
Hi!
Thanks Nicola for the reply. Can you provide your version of WikilinksNED Unseen-Mentions test file ? Also for generating Table 6 numbers you use which model (fairseq_entity_disambiguation_blink or fairseq_entity_disambiguation_aidayago ) ? also did you fine tune on WikilinksNED Unseen-Mentions train set ?
Unfortunately no. I have not access to the machines I used for these experiments.
For table 6 I think I used fairseq_entity_disambiguation_blink and I fine-tuned on the WikilinksNED Unseen-Mentions train set
Hi Nicola,
I have run the GENRE on WikilinksNED Unseen-Mentions dataset (taken from here "https://github.com/yasumasaonoe/ET4EL") and got an accuracy of 61.3, but the paper reports 63.52 (Table 6). I used the default settings: beam=10,max_len_a=384, max_len_b=15, and KILT trie. Can you tell me the reason for this difference? Is it because of not using "Yago trie"? Can you tell me while running your experiment on WikilinksNED Unseen-Mentions dataset did you lowercase all the sentences ?
Does GENRE output for entity disambiguation depends whether the context of the mention is written in lowercase or uppercase ? Please see the below output for sentence_1 and Sentence_2. Sentence_1 = ["Einstein was a [START_ENT] German [END_ENT] physicist."]
[{'text': 'Germans', 'score': tensor(-0.2991, device='cuda:0')}
Sentence_2 = ["einstein was a [START_ENT] german [END_ENT] physicist.”]
[{'text': 'Germany', 'score': tensor(-0.2907, device='cuda:0')},