Cargando…

Assessing the Quality of Mobile Health-Related Apps: Interrater Reliability Study of Two Guides

BACKGROUND: There is a huge number of health-related apps available, and the numbers are growing fast. However, many of them have been developed without any kind of quality control. In an attempt to contribute to the development of high-quality apps and enable existing apps to be assessed, several g...

Descripción completa

Detalles Bibliográficos
Autores principales: Miró, Jordi, Llorens-Vernet, Pere
Formato: Online Artículo Texto
Lenguaje:English
Publicado: JMIR Publications 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8094021/
https://www.ncbi.nlm.nih.gov/pubmed/33871376
http://dx.doi.org/10.2196/26471
_version_ 1783687936598867968
author Miró, Jordi
Llorens-Vernet, Pere
author_facet Miró, Jordi
Llorens-Vernet, Pere
author_sort Miró, Jordi
collection PubMed
description BACKGROUND: There is a huge number of health-related apps available, and the numbers are growing fast. However, many of them have been developed without any kind of quality control. In an attempt to contribute to the development of high-quality apps and enable existing apps to be assessed, several guides have been developed. OBJECTIVE: The main aim of this study was to study the interrater reliability of a new guide — the Mobile App Development and Assessment Guide (MAG) — and compare it with one of the most used guides in the field, the Mobile App Rating Scale (MARS). Moreover, we also focused on whether the interrater reliability of the measures is consistent across multiple types of apps and stakeholders. METHODS: In order to study the interrater reliability of the MAG and MARS, we evaluated the 4 most downloaded health apps for chronic health conditions in the medical category of IOS and Android devices (ie, App Store and Google Play). A group of 8 reviewers, representative of individuals that would be most knowledgeable and interested in the use and development of health-related apps and including different types of stakeholders such as clinical researchers, engineers, health care professionals, and end users as potential patients, independently evaluated the quality of the apps using the MAG and MARS. We calculated the Krippendorff alpha for every category in the 2 guides, for each type of reviewer and every app, separately and combined, to study the interrater reliability. RESULTS: Only a few categories of the MAG and MARS demonstrated a high interrater reliability. Although the MAG was found to be superior, there was considerable variation in the scores between the different types of reviewers. The categories with the highest interrater reliability in MAG were “Security” (α=0.78) and “Privacy” (α=0.73). In addition, 2 other categories, “Usability” and “Safety,” were very close to compliance (health care professionals: α=0.62 and 0.61, respectively). The total interrater reliability of the MAG (ie, for all categories) was 0.45, whereas the total interrater reliability of the MARS was 0.29. CONCLUSIONS: This study shows that some categories of MAG have significant interrater reliability. Importantly, the data show that the MAG scores are better than the ones provided by the MARS, which is the most commonly used guide in the area. However, there is great variability in the responses, which seems to be associated with subjective interpretation by the reviewers.
format Online
Article
Text
id pubmed-8094021
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher JMIR Publications
record_format MEDLINE/PubMed
spelling pubmed-80940212021-05-07 Assessing the Quality of Mobile Health-Related Apps: Interrater Reliability Study of Two Guides Miró, Jordi Llorens-Vernet, Pere JMIR Mhealth Uhealth Original Paper BACKGROUND: There is a huge number of health-related apps available, and the numbers are growing fast. However, many of them have been developed without any kind of quality control. In an attempt to contribute to the development of high-quality apps and enable existing apps to be assessed, several guides have been developed. OBJECTIVE: The main aim of this study was to study the interrater reliability of a new guide — the Mobile App Development and Assessment Guide (MAG) — and compare it with one of the most used guides in the field, the Mobile App Rating Scale (MARS). Moreover, we also focused on whether the interrater reliability of the measures is consistent across multiple types of apps and stakeholders. METHODS: In order to study the interrater reliability of the MAG and MARS, we evaluated the 4 most downloaded health apps for chronic health conditions in the medical category of IOS and Android devices (ie, App Store and Google Play). A group of 8 reviewers, representative of individuals that would be most knowledgeable and interested in the use and development of health-related apps and including different types of stakeholders such as clinical researchers, engineers, health care professionals, and end users as potential patients, independently evaluated the quality of the apps using the MAG and MARS. We calculated the Krippendorff alpha for every category in the 2 guides, for each type of reviewer and every app, separately and combined, to study the interrater reliability. RESULTS: Only a few categories of the MAG and MARS demonstrated a high interrater reliability. Although the MAG was found to be superior, there was considerable variation in the scores between the different types of reviewers. The categories with the highest interrater reliability in MAG were “Security” (α=0.78) and “Privacy” (α=0.73). In addition, 2 other categories, “Usability” and “Safety,” were very close to compliance (health care professionals: α=0.62 and 0.61, respectively). The total interrater reliability of the MAG (ie, for all categories) was 0.45, whereas the total interrater reliability of the MARS was 0.29. CONCLUSIONS: This study shows that some categories of MAG have significant interrater reliability. Importantly, the data show that the MAG scores are better than the ones provided by the MARS, which is the most commonly used guide in the area. However, there is great variability in the responses, which seems to be associated with subjective interpretation by the reviewers. JMIR Publications 2021-04-19 /pmc/articles/PMC8094021/ /pubmed/33871376 http://dx.doi.org/10.2196/26471 Text en ©Jordi Miró, Pere Llorens-Vernet. Originally published in JMIR mHealth and uHealth (https://mhealth.jmir.org), 19.04.2021. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR mHealth and uHealth, is properly cited. The complete bibliographic information, a link to the original publication on http://mhealth.jmir.org/, as well as this copyright and license information must be included.
spellingShingle Original Paper
Miró, Jordi
Llorens-Vernet, Pere
Assessing the Quality of Mobile Health-Related Apps: Interrater Reliability Study of Two Guides
title Assessing the Quality of Mobile Health-Related Apps: Interrater Reliability Study of Two Guides
title_full Assessing the Quality of Mobile Health-Related Apps: Interrater Reliability Study of Two Guides
title_fullStr Assessing the Quality of Mobile Health-Related Apps: Interrater Reliability Study of Two Guides
title_full_unstemmed Assessing the Quality of Mobile Health-Related Apps: Interrater Reliability Study of Two Guides
title_short Assessing the Quality of Mobile Health-Related Apps: Interrater Reliability Study of Two Guides
title_sort assessing the quality of mobile health-related apps: interrater reliability study of two guides
topic Original Paper
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8094021/
https://www.ncbi.nlm.nih.gov/pubmed/33871376
http://dx.doi.org/10.2196/26471
work_keys_str_mv AT mirojordi assessingthequalityofmobilehealthrelatedappsinterraterreliabilitystudyoftwoguides
AT llorensvernetpere assessingthequalityofmobilehealthrelatedappsinterraterreliabilitystudyoftwoguides