Cargando…

Tuning Fairness by Balancing Target Labels

The issue of fairness in machine learning models has recently attracted a lot of attention as ensuring it will ensure continued confidence of the general public in the deployment of machine learning systems. We focus on mitigating the harm incurred by a biased machine learning system that offers bet...

Descripción completa

Detalles Bibliográficos
Autores principales: Kehrenberg, Thomas, Chen, Zexun, Quadrianto, Novi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7861271/
https://www.ncbi.nlm.nih.gov/pubmed/33733151
http://dx.doi.org/10.3389/frai.2020.00033
_version_ 1783647049979265024
author Kehrenberg, Thomas
Chen, Zexun
Quadrianto, Novi
author_facet Kehrenberg, Thomas
Chen, Zexun
Quadrianto, Novi
author_sort Kehrenberg, Thomas
collection PubMed
description The issue of fairness in machine learning models has recently attracted a lot of attention as ensuring it will ensure continued confidence of the general public in the deployment of machine learning systems. We focus on mitigating the harm incurred by a biased machine learning system that offers better outputs (e.g., loans, job interviews) for certain groups than for others. We show that bias in the output can naturally be controlled in probabilistic models by introducing a latent target output. This formulation has several advantages: first, it is a unified framework for several notions of group fairness such as Demographic Parity and Equality of Opportunity; second, it is expressed as a marginalization instead of a constrained problem; and third, it allows the encoding of our knowledge of what unbiased outputs should be. Practically, the second allows us to avoid unstable constrained optimization procedures and to reuse off-the-shelf toolboxes. The latter translates to the ability to control the level of fairness by directly varying fairness target rates. In contrast, existing approaches rely on intermediate, arguably unintuitive, control parameters such as covariance thresholds.
format Online
Article
Text
id pubmed-7861271
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-78612712021-03-16 Tuning Fairness by Balancing Target Labels Kehrenberg, Thomas Chen, Zexun Quadrianto, Novi Front Artif Intell Artificial Intelligence The issue of fairness in machine learning models has recently attracted a lot of attention as ensuring it will ensure continued confidence of the general public in the deployment of machine learning systems. We focus on mitigating the harm incurred by a biased machine learning system that offers better outputs (e.g., loans, job interviews) for certain groups than for others. We show that bias in the output can naturally be controlled in probabilistic models by introducing a latent target output. This formulation has several advantages: first, it is a unified framework for several notions of group fairness such as Demographic Parity and Equality of Opportunity; second, it is expressed as a marginalization instead of a constrained problem; and third, it allows the encoding of our knowledge of what unbiased outputs should be. Practically, the second allows us to avoid unstable constrained optimization procedures and to reuse off-the-shelf toolboxes. The latter translates to the ability to control the level of fairness by directly varying fairness target rates. In contrast, existing approaches rely on intermediate, arguably unintuitive, control parameters such as covariance thresholds. Frontiers Media S.A. 2020-05-12 /pmc/articles/PMC7861271/ /pubmed/33733151 http://dx.doi.org/10.3389/frai.2020.00033 Text en Copyright © 2020 Kehrenberg, Chen and Quadrianto. http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Kehrenberg, Thomas
Chen, Zexun
Quadrianto, Novi
Tuning Fairness by Balancing Target Labels
title Tuning Fairness by Balancing Target Labels
title_full Tuning Fairness by Balancing Target Labels
title_fullStr Tuning Fairness by Balancing Target Labels
title_full_unstemmed Tuning Fairness by Balancing Target Labels
title_short Tuning Fairness by Balancing Target Labels
title_sort tuning fairness by balancing target labels
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7861271/
https://www.ncbi.nlm.nih.gov/pubmed/33733151
http://dx.doi.org/10.3389/frai.2020.00033
work_keys_str_mv AT kehrenbergthomas tuningfairnessbybalancingtargetlabels
AT chenzexun tuningfairnessbybalancingtargetlabels
AT quadriantonovi tuningfairnessbybalancingtargetlabels