Cargando…

How Validation Methodology Influences Human Activity Recognition Mobile Systems

In this article, we introduce explainable methods to understand how Human Activity Recognition (HAR) mobile systems perform based on the chosen validation strategies. Our results introduce a new way to discover potential bias problems that overestimate the prediction accuracy of an algorithm because...

Descripción completa

Detalles Bibliográficos
Autores principales: Bragança, Hendrio, Colonna, Juan G., Oliveira, Horácio A. B. F., Souto, Eduardo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8954513/
https://www.ncbi.nlm.nih.gov/pubmed/35336529
http://dx.doi.org/10.3390/s22062360
_version_ 1784676111153627136
author Bragança, Hendrio
Colonna, Juan G.
Oliveira, Horácio A. B. F.
Souto, Eduardo
author_facet Bragança, Hendrio
Colonna, Juan G.
Oliveira, Horácio A. B. F.
Souto, Eduardo
author_sort Bragança, Hendrio
collection PubMed
description In this article, we introduce explainable methods to understand how Human Activity Recognition (HAR) mobile systems perform based on the chosen validation strategies. Our results introduce a new way to discover potential bias problems that overestimate the prediction accuracy of an algorithm because of the inappropriate choice of validation methodology. We show how the SHAP (Shapley additive explanations) framework, used in literature to explain the predictions of any machine learning model, presents itself as a tool that can provide graphical insights into how human activity recognition models achieve their results. Now it is possible to analyze which features are important to a HAR system in each validation methodology in a simplified way. We not only demonstrate that the validation procedure k-folds cross-validation (k-CV), used in most works to evaluate the expected error in a HAR system, can overestimate by about 13% the prediction accuracy in three public datasets but also choose a different feature set when compared with the universal model. Combining explainable methods with machine learning algorithms has the potential to help new researchers look inside the decisions of the machine learning algorithms, avoiding most times the overestimation of prediction accuracy, understanding relations between features, and finding bias before deploying the system in real-world scenarios.
format Online
Article
Text
id pubmed-8954513
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-89545132022-03-26 How Validation Methodology Influences Human Activity Recognition Mobile Systems Bragança, Hendrio Colonna, Juan G. Oliveira, Horácio A. B. F. Souto, Eduardo Sensors (Basel) Article In this article, we introduce explainable methods to understand how Human Activity Recognition (HAR) mobile systems perform based on the chosen validation strategies. Our results introduce a new way to discover potential bias problems that overestimate the prediction accuracy of an algorithm because of the inappropriate choice of validation methodology. We show how the SHAP (Shapley additive explanations) framework, used in literature to explain the predictions of any machine learning model, presents itself as a tool that can provide graphical insights into how human activity recognition models achieve their results. Now it is possible to analyze which features are important to a HAR system in each validation methodology in a simplified way. We not only demonstrate that the validation procedure k-folds cross-validation (k-CV), used in most works to evaluate the expected error in a HAR system, can overestimate by about 13% the prediction accuracy in three public datasets but also choose a different feature set when compared with the universal model. Combining explainable methods with machine learning algorithms has the potential to help new researchers look inside the decisions of the machine learning algorithms, avoiding most times the overestimation of prediction accuracy, understanding relations between features, and finding bias before deploying the system in real-world scenarios. MDPI 2022-03-18 /pmc/articles/PMC8954513/ /pubmed/35336529 http://dx.doi.org/10.3390/s22062360 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Bragança, Hendrio
Colonna, Juan G.
Oliveira, Horácio A. B. F.
Souto, Eduardo
How Validation Methodology Influences Human Activity Recognition Mobile Systems
title How Validation Methodology Influences Human Activity Recognition Mobile Systems
title_full How Validation Methodology Influences Human Activity Recognition Mobile Systems
title_fullStr How Validation Methodology Influences Human Activity Recognition Mobile Systems
title_full_unstemmed How Validation Methodology Influences Human Activity Recognition Mobile Systems
title_short How Validation Methodology Influences Human Activity Recognition Mobile Systems
title_sort how validation methodology influences human activity recognition mobile systems
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8954513/
https://www.ncbi.nlm.nih.gov/pubmed/35336529
http://dx.doi.org/10.3390/s22062360
work_keys_str_mv AT bragancahendrio howvalidationmethodologyinfluenceshumanactivityrecognitionmobilesystems
AT colonnajuang howvalidationmethodologyinfluenceshumanactivityrecognitionmobilesystems
AT oliveirahoracioabf howvalidationmethodologyinfluenceshumanactivityrecognitionmobilesystems
AT soutoeduardo howvalidationmethodologyinfluenceshumanactivityrecognitionmobilesystems