Cargando…
Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics
The harmful impacts of algorithmic decision systems have recently come into focus, with many examples of machine learning (ML) models amplifying societal biases. In this paper, we propose adapting income inequality metrics from economics to complement existing model-level fairness metrics, which foc...
Autores principales: | , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9403369/ https://www.ncbi.nlm.nih.gov/pubmed/36033598 http://dx.doi.org/10.1016/j.patter.2022.100568 |
_version_ | 1784773360465477632 |
---|---|
author | Lazovich, Tomo Belli, Luca Gonzales, Aaron Bower, Amanda Tantipongpipat, Uthaipon Lum, Kristian Huszár, Ferenc Chowdhury, Rumman |
author_facet | Lazovich, Tomo Belli, Luca Gonzales, Aaron Bower, Amanda Tantipongpipat, Uthaipon Lum, Kristian Huszár, Ferenc Chowdhury, Rumman |
author_sort | Lazovich, Tomo |
collection | PubMed |
description | The harmful impacts of algorithmic decision systems have recently come into focus, with many examples of machine learning (ML) models amplifying societal biases. In this paper, we propose adapting income inequality metrics from economics to complement existing model-level fairness metrics, which focus on intergroup differences of model performance. In particular, we evaluate their ability to measure disparities between exposures that individuals receive in a production recommendation system, the Twitter algorithmic timeline. We define desirable criteria for metrics to be used in an operational setting by ML practitioners. We characterize engagements with content on Twitter using these metrics and use the results to evaluate the metrics with respect to our criteria. We also show that we can use these metrics to identify content suggestion algorithms that contribute more strongly to skewed outcomes between users. Overall, we conclude that these metrics can be a useful tool for auditing algorithms in production settings. |
format | Online Article Text |
id | pubmed-9403369 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Elsevier |
record_format | MEDLINE/PubMed |
spelling | pubmed-94033692022-08-26 Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics Lazovich, Tomo Belli, Luca Gonzales, Aaron Bower, Amanda Tantipongpipat, Uthaipon Lum, Kristian Huszár, Ferenc Chowdhury, Rumman Patterns (N Y) Article The harmful impacts of algorithmic decision systems have recently come into focus, with many examples of machine learning (ML) models amplifying societal biases. In this paper, we propose adapting income inequality metrics from economics to complement existing model-level fairness metrics, which focus on intergroup differences of model performance. In particular, we evaluate their ability to measure disparities between exposures that individuals receive in a production recommendation system, the Twitter algorithmic timeline. We define desirable criteria for metrics to be used in an operational setting by ML practitioners. We characterize engagements with content on Twitter using these metrics and use the results to evaluate the metrics with respect to our criteria. We also show that we can use these metrics to identify content suggestion algorithms that contribute more strongly to skewed outcomes between users. Overall, we conclude that these metrics can be a useful tool for auditing algorithms in production settings. Elsevier 2022-08-12 /pmc/articles/PMC9403369/ /pubmed/36033598 http://dx.doi.org/10.1016/j.patter.2022.100568 Text en © 2022 The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). |
spellingShingle | Article Lazovich, Tomo Belli, Luca Gonzales, Aaron Bower, Amanda Tantipongpipat, Uthaipon Lum, Kristian Huszár, Ferenc Chowdhury, Rumman Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics |
title | Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics |
title_full | Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics |
title_fullStr | Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics |
title_full_unstemmed | Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics |
title_short | Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics |
title_sort | measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9403369/ https://www.ncbi.nlm.nih.gov/pubmed/36033598 http://dx.doi.org/10.1016/j.patter.2022.100568 |
work_keys_str_mv | AT lazovichtomo measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics AT belliluca measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics AT gonzalesaaron measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics AT boweramanda measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics AT tantipongpipatuthaipon measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics AT lumkristian measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics AT huszarferenc measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics AT chowdhuryrumman measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics |