Back to Search Start Over

Gender Bias Mitigation for Bangla Classification Tasks

Authors :
Joy, Sajib Kumar Saha
Mahy, Arman Hassan
Sultana, Meherin
Abha, Azizah Mamun
Ahmmed, MD Piyal
Dong, Yue
Shahariar, G M
Publication Year :
2024

Abstract

In this study, we investigate gender bias in Bangla pretrained language models, a largely under explored area in low-resource languages. To assess this bias, we applied gender-name swapping techniques to existing datasets, creating four manually annotated, task-specific datasets for sentiment analysis, toxicity detection, hate speech detection, and sarcasm detection. By altering names and gender-specific terms, we ensured these datasets were suitable for detecting and mitigating gender bias. We then proposed a joint loss optimization technique to mitigate gender bias across task-specific pretrained models. Our approach was evaluated against existing bias mitigation methods, with results showing that our technique not only effectively reduces bias but also maintains competitive accuracy compared to other baseline approaches. To promote further research, we have made both our implementation and datasets publicly available https://github.com/sajib-kumar/Gender-Bias-Mitigation-From-Bangla-PLM

Details

Database :
arXiv
Publication Type :
Report
Accession number :
edsarx.2411.10636
Document Type :
Working Paper