@inproceedings{93a4195596bb4e98b7a75510e6001b47,
title = "Learning from Others: Similarity-based Regularization for Mitigating Dataset Bias",
abstract = "Common methods for mitigating spurious correlations in natural language understanding (NLU) usually operate in the output space, encouraging a main model to behave differently from a bias model by down-weighing examples where the bias model is confident. While improving out-of-distribution (OOD) performance, it was recently observed that the internal representations of the presumably debiased models are actually more, rather than less biased. We propose SimReg, a new method for debiasing internal model components via similarity-based regularization, in representation space: We encourage the model to learn representations that are either similar to an unbiased model or different from a biased model. We experiment with three NLU tasks and different kinds of biases. We find that SimReg improves OOD performance, with little in-distribution degradation. Moreover, the representations learned by SimReg are less biased than in other methods.",
author = "Reda Igbaria and Yonatan Belinkov",
note = "Publisher Copyright: {\textcopyright} 2024 Association for Computational Linguistics.; 9th Workshop on Representation Learning for NLP, RepL4NLP 2024 at ACL 2024 ; Conference date: 15-08-2024",
year = "2024",
language = "الإنجليزيّة",
series = "ACL 2024 - 9th Workshop on Representation Learning for NLP, RepL4NLP 2024 - Proceedings of the Workshop",
pages = "37--50",
editor = "Chen Zhao and Marius Mosbach and Pepa Atanasova and Seraphina Goldfarb-Tarrent and Peter Hase and Arian Hosseini and Maha Elbayad and Sandro Pezzelle and Maximilian Mozes",
booktitle = "ACL 2024 - 9th Workshop on Representation Learning for NLP, RepL4NLP 2024 - Proceedings of the Workshop",
}