Leveraging Prototypical Representations for Mitigating Social Bias without Demographic Information

Shadi Iskander, Kira Radinsky, Yonatan Belinkov

Research output: Contribution to conferencePaperpeer-review

Abstract

Mitigating social biases typically requires identifying the social groups associated with each data sample. In this paper, we present DAFAIR, a novel approach to address social bias in language models. Unlike traditional methods that rely on explicit demographic labels, our approach does not require any such information. Instead, we leverage predefined prototypical demographic texts and incorporate a regularization term during the fine-tuning process to mitigate bias in the model’s representations. Our empirical results across two tasks and two models demonstrate the effectiveness of our method compared to previous approaches that do not rely on labeled data. Moreover, with limited demographic-annotated data, our approach outperforms common debiasing approaches.

Original languageEnglish
Pages379-390
Number of pages12
DOIs
StatePublished - 2024
Event2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2024 - Hybrid, Mexico City, Mexico
Duration: 16 Jun 202421 Jun 2024

Conference

Conference2024 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL 2024
Country/TerritoryMexico
CityHybrid, Mexico City
Period16/06/2421/06/24

All Science Journal Classification (ASJC) codes

  • Computer Networks and Communications
  • Hardware and Architecture
  • Information Systems
  • Software

Fingerprint

Dive into the research topics of 'Leveraging Prototypical Representations for Mitigating Social Bias without Demographic Information'. Together they form a unique fingerprint.

Cite this