Cargando…

Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study

BACKGROUND: Approximately 1 in 5 American adults experience mental illness every year. Thus, mobile phone–based mental health prediction apps that use phone data and artificial intelligence techniques for mental health assessment have become increasingly important and are being rapidly developed. At...

Descripción completa

Detalles Bibliográficos
Autores principales: Park, Jinkyung, Arunachalam, Ramanathan, Silenzio, Vincent, Singh, Vivek K
Formato: Online Artículo Texto
Lenguaje:English
Publicado: JMIR Publications 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9240929/
https://www.ncbi.nlm.nih.gov/pubmed/35699997
http://dx.doi.org/10.2196/34366
_version_ 1784737675065950208
author Park, Jinkyung
Arunachalam, Ramanathan
Silenzio, Vincent
Singh, Vivek K
author_facet Park, Jinkyung
Arunachalam, Ramanathan
Silenzio, Vincent
Singh, Vivek K
author_sort Park, Jinkyung
collection PubMed
description BACKGROUND: Approximately 1 in 5 American adults experience mental illness every year. Thus, mobile phone–based mental health prediction apps that use phone data and artificial intelligence techniques for mental health assessment have become increasingly important and are being rapidly developed. At the same time, multiple artificial intelligence–related technologies (eg, face recognition and search results) have recently been reported to be biased regarding age, gender, and race. This study moves this discussion to a new domain: phone-based mental health assessment algorithms. It is important to ensure that such algorithms do not contribute to gender disparities through biased predictions across gender groups. OBJECTIVE: This research aimed to analyze the susceptibility of multiple commonly used machine learning approaches for gender bias in mobile mental health assessment and explore the use of an algorithmic disparate impact remover (DIR) approach to reduce bias levels while maintaining high accuracy. METHODS: First, we performed preprocessing and model training using the data set (N=55) obtained from a previous study. Accuracy levels and differences in accuracy across genders were computed using 5 different machine learning models. We selected the random forest model, which yielded the highest accuracy, for a more detailed audit and computed multiple metrics that are commonly used for fairness in the machine learning literature. Finally, we applied the DIR approach to reduce bias in the mental health assessment algorithm. RESULTS: The highest observed accuracy for the mental health assessment was 78.57%. Although this accuracy level raises optimism, the audit based on gender revealed that the performance of the algorithm was statistically significantly different between the male and female groups (eg, difference in accuracy across genders was 15.85%; P<.001). Similar trends were obtained for other fairness metrics. This disparity in performance was found to reduce significantly after the application of the DIR approach by adapting the data used for modeling (eg, the difference in accuracy across genders was 1.66%, and the reduction is statistically significant with P<.001). CONCLUSIONS: This study grounds the need for algorithmic auditing in phone-based mental health assessment algorithms and the use of gender as a protected attribute to study fairness in such settings. Such audits and remedial steps are the building blocks for the widespread adoption of fair and accurate mental health assessment algorithms in the future.
format Online
Article
Text
id pubmed-9240929
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher JMIR Publications
record_format MEDLINE/PubMed
spelling pubmed-92409292022-06-30 Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study Park, Jinkyung Arunachalam, Ramanathan Silenzio, Vincent Singh, Vivek K JMIR Form Res Original Paper BACKGROUND: Approximately 1 in 5 American adults experience mental illness every year. Thus, mobile phone–based mental health prediction apps that use phone data and artificial intelligence techniques for mental health assessment have become increasingly important and are being rapidly developed. At the same time, multiple artificial intelligence–related technologies (eg, face recognition and search results) have recently been reported to be biased regarding age, gender, and race. This study moves this discussion to a new domain: phone-based mental health assessment algorithms. It is important to ensure that such algorithms do not contribute to gender disparities through biased predictions across gender groups. OBJECTIVE: This research aimed to analyze the susceptibility of multiple commonly used machine learning approaches for gender bias in mobile mental health assessment and explore the use of an algorithmic disparate impact remover (DIR) approach to reduce bias levels while maintaining high accuracy. METHODS: First, we performed preprocessing and model training using the data set (N=55) obtained from a previous study. Accuracy levels and differences in accuracy across genders were computed using 5 different machine learning models. We selected the random forest model, which yielded the highest accuracy, for a more detailed audit and computed multiple metrics that are commonly used for fairness in the machine learning literature. Finally, we applied the DIR approach to reduce bias in the mental health assessment algorithm. RESULTS: The highest observed accuracy for the mental health assessment was 78.57%. Although this accuracy level raises optimism, the audit based on gender revealed that the performance of the algorithm was statistically significantly different between the male and female groups (eg, difference in accuracy across genders was 15.85%; P<.001). Similar trends were obtained for other fairness metrics. This disparity in performance was found to reduce significantly after the application of the DIR approach by adapting the data used for modeling (eg, the difference in accuracy across genders was 1.66%, and the reduction is statistically significant with P<.001). CONCLUSIONS: This study grounds the need for algorithmic auditing in phone-based mental health assessment algorithms and the use of gender as a protected attribute to study fairness in such settings. Such audits and remedial steps are the building blocks for the widespread adoption of fair and accurate mental health assessment algorithms in the future. JMIR Publications 2022-06-14 /pmc/articles/PMC9240929/ /pubmed/35699997 http://dx.doi.org/10.2196/34366 Text en ©Jinkyung Park, Ramanathan Arunachalam, Vincent Silenzio, Vivek K Singh. Originally published in JMIR Formative Research (https://formative.jmir.org), 14.06.2022. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included.
spellingShingle Original Paper
Park, Jinkyung
Arunachalam, Ramanathan
Silenzio, Vincent
Singh, Vivek K
Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study
title Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study
title_full Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study
title_fullStr Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study
title_full_unstemmed Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study
title_short Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study
title_sort fairness in mobile phone–based mental health assessment algorithms: exploratory study
topic Original Paper
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9240929/
https://www.ncbi.nlm.nih.gov/pubmed/35699997
http://dx.doi.org/10.2196/34366
work_keys_str_mv AT parkjinkyung fairnessinmobilephonebasedmentalhealthassessmentalgorithmsexploratorystudy
AT arunachalamramanathan fairnessinmobilephonebasedmentalhealthassessmentalgorithmsexploratorystudy
AT silenziovincent fairnessinmobilephonebasedmentalhealthassessmentalgorithmsexploratorystudy
AT singhvivekk fairnessinmobilephonebasedmentalhealthassessmentalgorithmsexploratorystudy