Cargando…
Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study
BACKGROUND: Approximately 1 in 5 American adults experience mental illness every year. Thus, mobile phone–based mental health prediction apps that use phone data and artificial intelligence techniques for mental health assessment have become increasingly important and are being rapidly developed. At...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
JMIR Publications
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9240929/ https://www.ncbi.nlm.nih.gov/pubmed/35699997 http://dx.doi.org/10.2196/34366 |
_version_ | 1784737675065950208 |
---|---|
author | Park, Jinkyung Arunachalam, Ramanathan Silenzio, Vincent Singh, Vivek K |
author_facet | Park, Jinkyung Arunachalam, Ramanathan Silenzio, Vincent Singh, Vivek K |
author_sort | Park, Jinkyung |
collection | PubMed |
description | BACKGROUND: Approximately 1 in 5 American adults experience mental illness every year. Thus, mobile phone–based mental health prediction apps that use phone data and artificial intelligence techniques for mental health assessment have become increasingly important and are being rapidly developed. At the same time, multiple artificial intelligence–related technologies (eg, face recognition and search results) have recently been reported to be biased regarding age, gender, and race. This study moves this discussion to a new domain: phone-based mental health assessment algorithms. It is important to ensure that such algorithms do not contribute to gender disparities through biased predictions across gender groups. OBJECTIVE: This research aimed to analyze the susceptibility of multiple commonly used machine learning approaches for gender bias in mobile mental health assessment and explore the use of an algorithmic disparate impact remover (DIR) approach to reduce bias levels while maintaining high accuracy. METHODS: First, we performed preprocessing and model training using the data set (N=55) obtained from a previous study. Accuracy levels and differences in accuracy across genders were computed using 5 different machine learning models. We selected the random forest model, which yielded the highest accuracy, for a more detailed audit and computed multiple metrics that are commonly used for fairness in the machine learning literature. Finally, we applied the DIR approach to reduce bias in the mental health assessment algorithm. RESULTS: The highest observed accuracy for the mental health assessment was 78.57%. Although this accuracy level raises optimism, the audit based on gender revealed that the performance of the algorithm was statistically significantly different between the male and female groups (eg, difference in accuracy across genders was 15.85%; P<.001). Similar trends were obtained for other fairness metrics. This disparity in performance was found to reduce significantly after the application of the DIR approach by adapting the data used for modeling (eg, the difference in accuracy across genders was 1.66%, and the reduction is statistically significant with P<.001). CONCLUSIONS: This study grounds the need for algorithmic auditing in phone-based mental health assessment algorithms and the use of gender as a protected attribute to study fairness in such settings. Such audits and remedial steps are the building blocks for the widespread adoption of fair and accurate mental health assessment algorithms in the future. |
format | Online Article Text |
id | pubmed-9240929 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | JMIR Publications |
record_format | MEDLINE/PubMed |
spelling | pubmed-92409292022-06-30 Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study Park, Jinkyung Arunachalam, Ramanathan Silenzio, Vincent Singh, Vivek K JMIR Form Res Original Paper BACKGROUND: Approximately 1 in 5 American adults experience mental illness every year. Thus, mobile phone–based mental health prediction apps that use phone data and artificial intelligence techniques for mental health assessment have become increasingly important and are being rapidly developed. At the same time, multiple artificial intelligence–related technologies (eg, face recognition and search results) have recently been reported to be biased regarding age, gender, and race. This study moves this discussion to a new domain: phone-based mental health assessment algorithms. It is important to ensure that such algorithms do not contribute to gender disparities through biased predictions across gender groups. OBJECTIVE: This research aimed to analyze the susceptibility of multiple commonly used machine learning approaches for gender bias in mobile mental health assessment and explore the use of an algorithmic disparate impact remover (DIR) approach to reduce bias levels while maintaining high accuracy. METHODS: First, we performed preprocessing and model training using the data set (N=55) obtained from a previous study. Accuracy levels and differences in accuracy across genders were computed using 5 different machine learning models. We selected the random forest model, which yielded the highest accuracy, for a more detailed audit and computed multiple metrics that are commonly used for fairness in the machine learning literature. Finally, we applied the DIR approach to reduce bias in the mental health assessment algorithm. RESULTS: The highest observed accuracy for the mental health assessment was 78.57%. Although this accuracy level raises optimism, the audit based on gender revealed that the performance of the algorithm was statistically significantly different between the male and female groups (eg, difference in accuracy across genders was 15.85%; P<.001). Similar trends were obtained for other fairness metrics. This disparity in performance was found to reduce significantly after the application of the DIR approach by adapting the data used for modeling (eg, the difference in accuracy across genders was 1.66%, and the reduction is statistically significant with P<.001). CONCLUSIONS: This study grounds the need for algorithmic auditing in phone-based mental health assessment algorithms and the use of gender as a protected attribute to study fairness in such settings. Such audits and remedial steps are the building blocks for the widespread adoption of fair and accurate mental health assessment algorithms in the future. JMIR Publications 2022-06-14 /pmc/articles/PMC9240929/ /pubmed/35699997 http://dx.doi.org/10.2196/34366 Text en ©Jinkyung Park, Ramanathan Arunachalam, Vincent Silenzio, Vivek K Singh. Originally published in JMIR Formative Research (https://formative.jmir.org), 14.06.2022. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Formative Research, is properly cited. The complete bibliographic information, a link to the original publication on https://formative.jmir.org, as well as this copyright and license information must be included. |
spellingShingle | Original Paper Park, Jinkyung Arunachalam, Ramanathan Silenzio, Vincent Singh, Vivek K Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study |
title | Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study |
title_full | Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study |
title_fullStr | Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study |
title_full_unstemmed | Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study |
title_short | Fairness in Mobile Phone–Based Mental Health Assessment Algorithms: Exploratory Study |
title_sort | fairness in mobile phone–based mental health assessment algorithms: exploratory study |
topic | Original Paper |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9240929/ https://www.ncbi.nlm.nih.gov/pubmed/35699997 http://dx.doi.org/10.2196/34366 |
work_keys_str_mv | AT parkjinkyung fairnessinmobilephonebasedmentalhealthassessmentalgorithmsexploratorystudy AT arunachalamramanathan fairnessinmobilephonebasedmentalhealthassessmentalgorithmsexploratorystudy AT silenziovincent fairnessinmobilephonebasedmentalhealthassessmentalgorithmsexploratorystudy AT singhvivekk fairnessinmobilephonebasedmentalhealthassessmentalgorithmsexploratorystudy |