Optimization of Bias Mitigation in Word Embeddings: A Methodological Approach

Zambrano M.J.; Bravo-Marquez F.

Keywords: Bias; Natural Language Processing; Word Embeddings

Abstract

Word embeddings (WEs) often reflect biases present in their training data, and various bias mitigation and evaluation techniques have been proposed to address this. Existing benchmarks for comparing different debiasing methods overlook two factors: the choice of training words and model hyper-parameters. We propose a robust comparison methodology that incorporates them using nested cross-validation, hyper-parameter optimization, and the corrected paired Student's t-test. Our results show that when using our evaluation approach many recent debiasing methods do not offer statistically significant improvements over the original hard debiasing model.

Más información

Título según SCOPUS: Optimization of Bias Mitigation in Word Embeddings: A Methodological Approach
Título de la Revista: Proceedings - 2024 IEEE/WIC International Conference on Web Intelligence and Intelligent Agent Technology, WI-IAT 2024
Editorial: Institute of Electrical and Electronics Engineers Inc.
Fecha de publicación: 2024
Página final: 484
Idioma: English
DOI:

10.1109/WI-IAT62293.2024.00077

Notas: SCOPUS