Closed miniweeds closed 1 year ago
Or is it because the jigsaw dataset in the paper is different from what I used? I tried the non-English jigsaw dataset too. The performance was worse.
I see the problem now. The moderation API only classifies the following categories: "hate”, “hate/threatening”, "self-harm”, "sexual”, "sexual/minors”, "violence”, “violence/graphic". The jigsaw dataset I used covers much more categories. That explains why the API got such low AUPRC, the model and the test set don't align. In this case this API is not suitable for the jigsaw type of problems.
Here are the categories in Jigsaw train dataset: severe_toxicity,obscene,identity_attack,insult,threat,asian,atheist,bisexual,black,buddhist,christian,female,heterosexual,hindu,homosexual_gay_or_lesbian,intellectual_or_learning_disability,jewish,latino,male,muslim,other_disability,other_gender,other_race_or_ethnicity,other_religion,other_sexual_orientation,physical_disability,psychiatric_or_mental_illness,transgender,white.
I tried to test the moderation API performance with the jigsaw dataset from Kaggle. The performance is quite worse than what was reported in the paper. Why? What am I missing? Here are the parameters for my test:
My Test result:
AUPRC: 0.33
Paper (https://arxiv.org/pdf/2208.03274.pdf) reported much better AUPRC, as below: Jigsaw Identity-hate .6890 Insult .8548 Obscene .8353 Threat .6144 Toxic .9304*