Cargando…

Fairness and Accuracy Under Domain Generalization

As machine learning (ML) algorithms are increasingly used in high-stakes applications, concerns have arisen that they may be biased against certain social groups. Although many approaches have been proposed to make ML models fair, they typically rely on the assumption that data distributions in trai...

Descripción completa

Detalles Bibliográficos
Autores principales: Pham, Thai-Hoang, Zhang, Xueru, Zhang, Ping
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Cornell University 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10246117/
https://www.ncbi.nlm.nih.gov/pubmed/37292471
_version_ 1785054981216272384
author Pham, Thai-Hoang
Zhang, Xueru
Zhang, Ping
author_facet Pham, Thai-Hoang
Zhang, Xueru
Zhang, Ping
author_sort Pham, Thai-Hoang
collection PubMed
description As machine learning (ML) algorithms are increasingly used in high-stakes applications, concerns have arisen that they may be biased against certain social groups. Although many approaches have been proposed to make ML models fair, they typically rely on the assumption that data distributions in training and deployment are identical. Unfortunately, this is commonly violated in practice and a model that is fair during training may lead to an unexpected outcome during its deployment. Although the problem of designing robust ML models under dataset shifts has been widely studied, most existing works focus only on the transfer of accuracy. In this paper, we study the transfer of both fairness and accuracy under domain generalization where the data at test time may be sampled from never-before-seen domains. We first develop theoretical bounds on the unfairness and expected loss at deployment, and then derive sufficient conditions under which fairness and accuracy can be perfectly transferred via invariant representation learning. Guided by this, we design a learning algorithm such that fair ML models learned with training data still have high fairness and accuracy when deployment environments change. Experiments on real-world data validate the proposed algorithm. Model implementation is available at https://github.com/pth1993/FATDM.
format Online
Article
Text
id pubmed-10246117
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Cornell University
record_format MEDLINE/PubMed
spelling pubmed-102461172023-06-08 Fairness and Accuracy Under Domain Generalization Pham, Thai-Hoang Zhang, Xueru Zhang, Ping ArXiv Article As machine learning (ML) algorithms are increasingly used in high-stakes applications, concerns have arisen that they may be biased against certain social groups. Although many approaches have been proposed to make ML models fair, they typically rely on the assumption that data distributions in training and deployment are identical. Unfortunately, this is commonly violated in practice and a model that is fair during training may lead to an unexpected outcome during its deployment. Although the problem of designing robust ML models under dataset shifts has been widely studied, most existing works focus only on the transfer of accuracy. In this paper, we study the transfer of both fairness and accuracy under domain generalization where the data at test time may be sampled from never-before-seen domains. We first develop theoretical bounds on the unfairness and expected loss at deployment, and then derive sufficient conditions under which fairness and accuracy can be perfectly transferred via invariant representation learning. Guided by this, we design a learning algorithm such that fair ML models learned with training data still have high fairness and accuracy when deployment environments change. Experiments on real-world data validate the proposed algorithm. Model implementation is available at https://github.com/pth1993/FATDM. Cornell University 2023-01-30 /pmc/articles/PMC10246117/ /pubmed/37292471 Text en https://creativecommons.org/licenses/by-nc-nd/4.0/This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License (https://creativecommons.org/licenses/by-nc-nd/4.0/) , which allows reusers to copy and distribute the material in any medium or format in unadapted form only, for noncommercial purposes only, and only so long as attribution is given to the creator.
spellingShingle Article
Pham, Thai-Hoang
Zhang, Xueru
Zhang, Ping
Fairness and Accuracy Under Domain Generalization
title Fairness and Accuracy Under Domain Generalization
title_full Fairness and Accuracy Under Domain Generalization
title_fullStr Fairness and Accuracy Under Domain Generalization
title_full_unstemmed Fairness and Accuracy Under Domain Generalization
title_short Fairness and Accuracy Under Domain Generalization
title_sort fairness and accuracy under domain generalization
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10246117/
https://www.ncbi.nlm.nih.gov/pubmed/37292471
work_keys_str_mv AT phamthaihoang fairnessandaccuracyunderdomaingeneralization
AT zhangxueru fairnessandaccuracyunderdomaingeneralization
AT zhangping fairnessandaccuracyunderdomaingeneralization