unitaryai / detoxify

Trained models & code to predict toxic comments on all 3 Jigsaw Toxic Comment Challenges. Built using ⚡ Pytorch Lightning and 🤗 Transformers. For access to our API, please email us at contact@unitary.ai.
https://www.unitary.ai/
Apache License 2.0
967 stars 114 forks source link

Issues when training unintended bias model #110

Closed MLRadfys closed 5 months ago

MLRadfys commented 5 months ago

Hi and thanks for this great repository!

I tried to fine tune the Roberta model on my own for unintended bias in comments classification. Unfortunately, all of the subgroup AUCs are close to 0.5.

I used the standard configuration given in this repo.

Does anyone tried fine tuning the models on their own?

I also evaluated the unbiased model given in the repo, leading to a similar result:

image

Thanks in advance,

/M

EDIT:

Hi again!

My fault --> I evaluated on public_test.csv and computed the bias AUCs on private_test.csv :-( Now it is working :-)

/M

jamt9000 commented 5 months ago

Glad you got it working!