Open pramitchoudhary opened 8 months ago
Hi @pramitchoudhary
Sorry for my late reply & thanks for looking into the annotation. Yes, it is possible that some sentences were annotated incorrectly. We were trying our best to ensure that annotations (using actual Wikipedia articles as ground-truth) are as accurate as possible.
Feel free to revise the annotations if you spot errors
Hi there, awesome stuff, thanks for sharing.
Question: Are there possible annotation errors in the eval dataset (_wiki_bio_gpt3hallucination)?
Example: Observing example 6, index 11 Example: 6
index =11
Ground Truth:
major_inaccurate
Should be:accurate
? Observation: During brief analysis, only NLI scores aggregated at the sentence level seem to agree with the ground truth? The code is provided below.Thoughts?