Cargando…

Measuring disparate outcomes of content recommendation algorithms with distributional inequality metrics

The harmful impacts of algorithmic decision systems have recently come into focus, with many examples of machine learning (ML) models amplifying societal biases. In this paper, we propose adapting income inequality metrics from economics to complement existing model-level fairness metrics, which foc...

Descripción completa

Detalles Bibliográficos
Autores principales: Lazovich, Tomo, Belli, Luca, Gonzales, Aaron, Bower, Amanda, Tantipongpipat, Uthaipon, Lum, Kristian, Huszár, Ferenc, Chowdhury, Rumman
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9403369/
https://www.ncbi.nlm.nih.gov/pubmed/36033598
http://dx.doi.org/10.1016/j.patter.2022.100568
Descripción
Sumario:The harmful impacts of algorithmic decision systems have recently come into focus, with many examples of machine learning (ML) models amplifying societal biases. In this paper, we propose adapting income inequality metrics from economics to complement existing model-level fairness metrics, which focus on intergroup differences of model performance. In particular, we evaluate their ability to measure disparities between exposures that individuals receive in a production recommendation system, the Twitter algorithmic timeline. We define desirable criteria for metrics to be used in an operational setting by ML practitioners. We characterize engagements with content on Twitter using these metrics and use the results to evaluate the metrics with respect to our criteria. We also show that we can use these metrics to identify content suggestion algorithms that contribute more strongly to skewed outcomes between users. Overall, we conclude that these metrics can be a useful tool for auditing algorithms in production settings.