Addressing Healthcare-related Racial and LGBTQ+ Biases in Pretrained Language Models

Link:

https://aclanthology.org/2024.findings-naacl.278/

Title:

Addressing Healthcare-related Racial and LGBTQ+ Biases in Pretrained Language Models

Abstract:

Recent studies have highlighted the issue of Pretrained Language Models (PLMs) inadvertently propagating social stigmas and stereotypes, a critical concern given their widespread use. This is particularly problematic in sensitive areas like healthcare, where such biases could lead to detrimental outcomes. Our research addresses this by adapting two intrinsic bias benchmarks to quantify racial and LGBTQ+ biases in prevalent PLMs. We also empirically evaluate the effectiveness of various debiasing methods in mitigating these biases. Furthermore, we assess the impact of debiasing on both Natural Language Understanding and specific biomedical applications. Our findings reveal that while PLMs commonly exhibit healthcare-related racial and LGBTQ+ biases, the applied debiasing techniques successfully reduce these biases without compromising the models’ performance in downstream tasks

Citation:

Xie S, Hassanpour S, Vosoughi S. Addressing Healthcare-related Racial and LGBTQ+ Biases in Pretrained Language Models. InFindings of the Association for Computational Linguistics: NAACL 2024, 2024 Jun (pp. 4451-4464)

Previous
Previous

Deep Learning for Grading Endometrial Cancer

Next
Next

Masked pre-training of transformers for histology image analysis