Mitigating Extrinsic Gender Bias for Bangla Classification Tasks

📰 ArXiv cs.AI

arXiv:2411.10636v2 Announce Type: replace-cross Abstract: In this study, we investigate extrinsic gender bias in Bangla pretrained language models, a largely underexplored area in low-resource languages. To assess this bias, we construct four manually annotated, task-specific benchmark datasets for sentiment analysis, toxicity detection, hate speech detection, and sarcasm detection. Each dataset is augmented using nuanced gender perturbations, where we systematically swap gendered names and term

Published 13 Apr 2026
Read full paper → ← Back to Reads