During the last years, the phenomenon of hate against women increased exponentially especially in online environments such as microblogs. Although this alarming phenomenon has triggered many studies both from computational linguistic and machine learning points of view, less effort has been spent to analyze if those misogyny detection models are affected by an unintended bias. This can lead the models to associate unreasonably high misogynous scores to a non-misogynous text only because it contains certain terms, called identity terms. This work is the first attempt to address the problem of measuring and mitigating unintended bias in machine learning models trained for the misogyny detection task. We propose a novel synthetic test set that can be used as evaluation framework for measuring the unintended bias and different mitigation strategies specific for this task. Moreover, we provide a misogyny detection model that demonstrate to obtain the best classification performance in the state-of-the-art. Experimental results on recently introduced bias metrics confirm the ability of the bias mitigation treatment to reduce the unintended bias of the proposed misogyny detection model.

Nozza, D., Volpetti, C., Fersini, E. (2019). Unintended bias in misogyny detection. In Proceedings - 2019 IEEE/WIC/ACM International Conference on Web Intelligence, WI 2019 (pp.149-155). New York, NY, USA : Association for Computing Machinery, Inc [10.1145/3350546.3352512].

Unintended bias in misogyny detection

Nozza, D
;
Fersini E
2019

Abstract

During the last years, the phenomenon of hate against women increased exponentially especially in online environments such as microblogs. Although this alarming phenomenon has triggered many studies both from computational linguistic and machine learning points of view, less effort has been spent to analyze if those misogyny detection models are affected by an unintended bias. This can lead the models to associate unreasonably high misogynous scores to a non-misogynous text only because it contains certain terms, called identity terms. This work is the first attempt to address the problem of measuring and mitigating unintended bias in machine learning models trained for the misogyny detection task. We propose a novel synthetic test set that can be used as evaluation framework for measuring the unintended bias and different mitigation strategies specific for this task. Moreover, we provide a misogyny detection model that demonstrate to obtain the best classification performance in the state-of-the-art. Experimental results on recently introduced bias metrics confirm the ability of the bias mitigation treatment to reduce the unintended bias of the proposed misogyny detection model.
paper
Bias measuring; Bias mitigation; Deep learning; Misogyny detection
English
19th IEEE/WIC/ACM International Conference on Web Intelligence, WI 2019 OCT 13-17
2019
Barnaghi, P; Gottlob, G; Manolopoulos, Y; Tzouramanis, T; Vakali, A
Proceedings - 2019 IEEE/WIC/ACM International Conference on Web Intelligence, WI 2019
9781450369343
2019
149
155
https://dl.acm.org/doi/10.1145/3350546.3352512
none
Nozza, D., Volpetti, C., Fersini, E. (2019). Unintended bias in misogyny detection. In Proceedings - 2019 IEEE/WIC/ACM International Conference on Web Intelligence, WI 2019 (pp.149-155). New York, NY, USA : Association for Computing Machinery, Inc [10.1145/3350546.3352512].
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/10281/256743
Citazioni
  • Scopus 62
  • ???jsp.display-item.citation.isi??? 31
Social impact