TY - JOUR
T1 - An intentional approach to managing bias in general purpose embedding models
AU - Weng, Wei Hung
AU - Sellergen, Andrew
AU - Kiraly, Atilla P.
AU - D'Amour, Alexander
AU - Park, Jungyeon
AU - Pilgrim, Rory
AU - Pfohl, Stephen
AU - Lau, Charles
AU - Natarajan, Vivek
AU - Azizi, Shekoofeh
AU - Karthikesalingam, Alan
AU - Cole-Lewis, Heather
AU - Matias, Yossi
AU - Corrado, Greg S.
AU - Webster, Dale R.
AU - Shetty, Shravya
AU - Prabhakara, Shruthi
AU - Eswaran, Krish
AU - Celi, Leo A.G.
AU - Liu, Yun
N1 - Publisher Copyright: © 2024 The Author(s). Published by Elsevier Ltd. This is an Open Access article under the CC BY 4.0 license
PY - 2024/2
Y1 - 2024/2
N2 - Advances in machine learning for health care have brought concerns about bias from the research community; specifically, the introduction, perpetuation, or exacerbation of care disparities. Reinforcing these concerns is the finding that medical images often reveal signals about sensitive attributes in ways that are hard to pinpoint by both algorithms and people. This finding raises a question about how to best design general purpose pretrained embeddings (GPPEs, defined as embeddings meant to support a broad array of use cases) for building downstream models that are free from particular types of bias. The downstream model should be carefully evaluated for bias, and audited and improved as appropriate. However, in our view, well intentioned attempts to prevent the upstream components—GPPEs—from learning sensitive attributes can have unintended consequences on the downstream models. Despite producing a veneer of technical neutrality, the resultant end-to-end system might still be biased or poorly performing. We present reasons, by building on previously published data, to support the reasoning that GPPEs should ideally contain as much information as the original data contain, and highlight the perils of trying to remove sensitive attributes from a GPPE. We also emphasise that downstream prediction models trained for specific tasks and settings, whether developed using GPPEs or not, should be carefully designed and evaluated to avoid bias that makes models vulnerable to issues such as distributional shift. These evaluations should be done by a diverse team, including social scientists, on a diverse cohort representing the full breadth of the patient population for which the final model is intended.
AB - Advances in machine learning for health care have brought concerns about bias from the research community; specifically, the introduction, perpetuation, or exacerbation of care disparities. Reinforcing these concerns is the finding that medical images often reveal signals about sensitive attributes in ways that are hard to pinpoint by both algorithms and people. This finding raises a question about how to best design general purpose pretrained embeddings (GPPEs, defined as embeddings meant to support a broad array of use cases) for building downstream models that are free from particular types of bias. The downstream model should be carefully evaluated for bias, and audited and improved as appropriate. However, in our view, well intentioned attempts to prevent the upstream components—GPPEs—from learning sensitive attributes can have unintended consequences on the downstream models. Despite producing a veneer of technical neutrality, the resultant end-to-end system might still be biased or poorly performing. We present reasons, by building on previously published data, to support the reasoning that GPPEs should ideally contain as much information as the original data contain, and highlight the perils of trying to remove sensitive attributes from a GPPE. We also emphasise that downstream prediction models trained for specific tasks and settings, whether developed using GPPEs or not, should be carefully designed and evaluated to avoid bias that makes models vulnerable to issues such as distributional shift. These evaluations should be done by a diverse team, including social scientists, on a diverse cohort representing the full breadth of the patient population for which the final model is intended.
UR - http://www.scopus.com/inward/record.url?scp=85182977540&partnerID=8YFLogxK
U2 - 10.1016/S2589-7500(23)00227-3
DO - 10.1016/S2589-7500(23)00227-3
M3 - مقالة مرجعية
C2 - 38278614
SN - 2589-7500
VL - 6
SP - e126-e130
JO - The Lancet Digital Health
JF - The Lancet Digital Health
IS - 2
ER -