Cargando…

Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics

The harmful impacts of algorithmic decision systems have recently come into focus, with many examples of machine learning (ML) models amplifying societal biases. In this paper, we propose adapting income inequality metrics from economics to complement existing model-level fairness metrics, which foc...

Descripción completa

Detalles Bibliográficos
Autores principales: Lazovich, Tomo, Belli, Luca, Gonzales, Aaron, Bower, Amanda, Tantipongpipat, Uthaipon, Lum, Kristian, Huszár, Ferenc, Chowdhury, Rumman
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9403369/
https://www.ncbi.nlm.nih.gov/pubmed/36033598
http://dx.doi.org/10.1016/j.patter.2022.100568
_version_ 1784773360465477632
author Lazovich, Tomo
Belli, Luca
Gonzales, Aaron
Bower, Amanda
Tantipongpipat, Uthaipon
Lum, Kristian
Huszár, Ferenc
Chowdhury, Rumman
author_facet Lazovich, Tomo
Belli, Luca
Gonzales, Aaron
Bower, Amanda
Tantipongpipat, Uthaipon
Lum, Kristian
Huszár, Ferenc
Chowdhury, Rumman
author_sort Lazovich, Tomo
collection PubMed
description The harmful impacts of algorithmic decision systems have recently come into focus, with many examples of machine learning (ML) models amplifying societal biases. In this paper, we propose adapting income inequality metrics from economics to complement existing model-level fairness metrics, which focus on intergroup differences of model performance. In particular, we evaluate their ability to measure disparities between exposures that individuals receive in a production recommendation system, the Twitter algorithmic timeline. We define desirable criteria for metrics to be used in an operational setting by ML practitioners. We characterize engagements with content on Twitter using these metrics and use the results to evaluate the metrics with respect to our criteria. We also show that we can use these metrics to identify content suggestion algorithms that contribute more strongly to skewed outcomes between users. Overall, we conclude that these metrics can be a useful tool for auditing algorithms in production settings.
format Online
Article
Text
id pubmed-9403369
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Elsevier
record_format MEDLINE/PubMed
spelling pubmed-94033692022-08-26 Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics Lazovich, Tomo Belli, Luca Gonzales, Aaron Bower, Amanda Tantipongpipat, Uthaipon Lum, Kristian Huszár, Ferenc Chowdhury, Rumman Patterns (N Y) Article The harmful impacts of algorithmic decision systems have recently come into focus, with many examples of machine learning (ML) models amplifying societal biases. In this paper, we propose adapting income inequality metrics from economics to complement existing model-level fairness metrics, which focus on intergroup differences of model performance. In particular, we evaluate their ability to measure disparities between exposures that individuals receive in a production recommendation system, the Twitter algorithmic timeline. We define desirable criteria for metrics to be used in an operational setting by ML practitioners. We characterize engagements with content on Twitter using these metrics and use the results to evaluate the metrics with respect to our criteria. We also show that we can use these metrics to identify content suggestion algorithms that contribute more strongly to skewed outcomes between users. Overall, we conclude that these metrics can be a useful tool for auditing algorithms in production settings. Elsevier 2022-08-12 /pmc/articles/PMC9403369/ /pubmed/36033598 http://dx.doi.org/10.1016/j.patter.2022.100568 Text en © 2022 The Authors https://creativecommons.org/licenses/by-nc-nd/4.0/This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
spellingShingle Article
Lazovich, Tomo
Belli, Luca
Gonzales, Aaron
Bower, Amanda
Tantipongpipat, Uthaipon
Lum, Kristian
Huszár, Ferenc
Chowdhury, Rumman
Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics
title Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics
title_full Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics
title_fullStr Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics
title_full_unstemmed Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics
title_short Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics
title_sort measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9403369/
https://www.ncbi.nlm.nih.gov/pubmed/36033598
http://dx.doi.org/10.1016/j.patter.2022.100568
work_keys_str_mv AT lazovichtomo measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics
AT belliluca measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics
AT gonzalesaaron measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics
AT boweramanda measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics
AT tantipongpipatuthaipon measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics
AT lumkristian measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics
AT huszarferenc measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics
AT chowdhuryrumman measuringdisparateoutcomesofcontentrecommendationalgorithmswithdistributionalinequalitymetrics