Cargando…
Interventional Fairness with Indirect Knowledge of Unobserved Protected Attributes
The deployment of machine learning (ML) systems in applications with societal impact has motivated the study of fairness for marginalized groups. Often, the protected attribute is absent from the training dataset for legal reasons. However, datasets still contain proxy attributes that capture protec...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8699829/ https://www.ncbi.nlm.nih.gov/pubmed/34945877 http://dx.doi.org/10.3390/e23121571 |
_version_ | 1784620608210862080 |
---|---|
author | Galhotra, Sainyam Shanmugam, Karthikeyan Sattigeri, Prasanna Varshney, Kush R. |
author_facet | Galhotra, Sainyam Shanmugam, Karthikeyan Sattigeri, Prasanna Varshney, Kush R. |
author_sort | Galhotra, Sainyam |
collection | PubMed |
description | The deployment of machine learning (ML) systems in applications with societal impact has motivated the study of fairness for marginalized groups. Often, the protected attribute is absent from the training dataset for legal reasons. However, datasets still contain proxy attributes that capture protected information and can inject unfairness in the ML model. Some deployed systems allow auditors, decision makers, or affected users to report issues or seek recourse by flagging individual samples. In this work, we examine such systems and consider a feedback-based framework where the protected attribute is unavailable and the flagged samples are indirect knowledge. The reported samples are used as guidance to identify the proxy attributes that are causally dependent on the (unknown) protected attribute. We work under the causal interventional fairness paradigm. Without requiring the underlying structural causal model a priori, we propose an approach that performs conditional independence tests on observed data to identify such proxy attributes. We theoretically prove the optimality of our algorithm, bound its complexity, and complement it with an empirical evaluation demonstrating its efficacy on various real-world and synthetic datasets. |
format | Online Article Text |
id | pubmed-8699829 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-86998292021-12-24 Interventional Fairness with Indirect Knowledge of Unobserved Protected Attributes Galhotra, Sainyam Shanmugam, Karthikeyan Sattigeri, Prasanna Varshney, Kush R. Entropy (Basel) Article The deployment of machine learning (ML) systems in applications with societal impact has motivated the study of fairness for marginalized groups. Often, the protected attribute is absent from the training dataset for legal reasons. However, datasets still contain proxy attributes that capture protected information and can inject unfairness in the ML model. Some deployed systems allow auditors, decision makers, or affected users to report issues or seek recourse by flagging individual samples. In this work, we examine such systems and consider a feedback-based framework where the protected attribute is unavailable and the flagged samples are indirect knowledge. The reported samples are used as guidance to identify the proxy attributes that are causally dependent on the (unknown) protected attribute. We work under the causal interventional fairness paradigm. Without requiring the underlying structural causal model a priori, we propose an approach that performs conditional independence tests on observed data to identify such proxy attributes. We theoretically prove the optimality of our algorithm, bound its complexity, and complement it with an empirical evaluation demonstrating its efficacy on various real-world and synthetic datasets. MDPI 2021-11-25 /pmc/articles/PMC8699829/ /pubmed/34945877 http://dx.doi.org/10.3390/e23121571 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Galhotra, Sainyam Shanmugam, Karthikeyan Sattigeri, Prasanna Varshney, Kush R. Interventional Fairness with Indirect Knowledge of Unobserved Protected Attributes |
title | Interventional Fairness with Indirect Knowledge of Unobserved Protected Attributes |
title_full | Interventional Fairness with Indirect Knowledge of Unobserved Protected Attributes |
title_fullStr | Interventional Fairness with Indirect Knowledge of Unobserved Protected Attributes |
title_full_unstemmed | Interventional Fairness with Indirect Knowledge of Unobserved Protected Attributes |
title_short | Interventional Fairness with Indirect Knowledge of Unobserved Protected Attributes |
title_sort | interventional fairness with indirect knowledge of unobserved protected attributes |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8699829/ https://www.ncbi.nlm.nih.gov/pubmed/34945877 http://dx.doi.org/10.3390/e23121571 |
work_keys_str_mv | AT galhotrasainyam interventionalfairnesswithindirectknowledgeofunobservedprotectedattributes AT shanmugamkarthikeyan interventionalfairnesswithindirectknowledgeofunobservedprotectedattributes AT sattigeriprasanna interventionalfairnesswithindirectknowledgeofunobservedprotectedattributes AT varshneykushr interventionalfairnesswithindirectknowledgeofunobservedprotectedattributes |