TY - GEN
T1 - Evaluating Gender Bias in Hindi-English Machine Translation
AU - Gupta, Gauri
AU - Ramesh, Krithika
AU - Singh, Sanjay
N1 - Publisher Copyright:
©2021 Association for Computational Linguistics.
PY - 2021
Y1 - 2021
N2 - With language models being deployed increasingly in the real world, it is essential to address the issue of the fairness of their outputs. The word embedding representations of these language models often implicitly draw unwanted associations that form a social bias within the model. The nature of gendered languages like Hindi, poses an additional problem to the quantification and mitigation of bias, owing to the change in the form of the words in the sentence, based on the gender of the subject. Additionally, there is sparse work done in the realm of measuring and debiasing systems for Indic languages. In our work, we attempt to evaluate and quantify the gender bias within a Hindi-English machine translation system.We implement a modified version of the existing TGBI metric based on the grammatical considerations for Hindi. We also compare and contrast the resulting bias measurements across multiple metrics for pre-trained embeddings and the ones learned by our machine translation model.
AB - With language models being deployed increasingly in the real world, it is essential to address the issue of the fairness of their outputs. The word embedding representations of these language models often implicitly draw unwanted associations that form a social bias within the model. The nature of gendered languages like Hindi, poses an additional problem to the quantification and mitigation of bias, owing to the change in the form of the words in the sentence, based on the gender of the subject. Additionally, there is sparse work done in the realm of measuring and debiasing systems for Indic languages. In our work, we attempt to evaluate and quantify the gender bias within a Hindi-English machine translation system.We implement a modified version of the existing TGBI metric based on the grammatical considerations for Hindi. We also compare and contrast the resulting bias measurements across multiple metrics for pre-trained embeddings and the ones learned by our machine translation model.
UR - http://www.scopus.com/inward/record.url?scp=85122052602&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85122052602&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85122052602
T3 - GeBNLP 2021 - 3rd Workshop on Gender Bias in Natural Language Processing, Proceedings
SP - 16
EP - 23
BT - GeBNLP 2021 - 3rd Workshop on Gender Bias in Natural Language Processing, Proceedings
A2 - Costa-jussa, Marta R.
A2 - Gonen, Hila
A2 - Hardmeier, Christian
A2 - Hardmeier, Christian
A2 - Webster, Kellie
PB - Association for Computational Linguistics (ACL)
T2 - 3rd Workshop on Gender Bias in Natural Language Processing, GeBNLP 2021
Y2 - 5 August 2021
ER -