Cargando…
How Validation Methodology Influences Human Activity Recognition Mobile Systems
In this article, we introduce explainable methods to understand how Human Activity Recognition (HAR) mobile systems perform based on the chosen validation strategies. Our results introduce a new way to discover potential bias problems that overestimate the prediction accuracy of an algorithm because...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8954513/ https://www.ncbi.nlm.nih.gov/pubmed/35336529 http://dx.doi.org/10.3390/s22062360 |
_version_ | 1784676111153627136 |
---|---|
author | Bragança, Hendrio Colonna, Juan G. Oliveira, Horácio A. B. F. Souto, Eduardo |
author_facet | Bragança, Hendrio Colonna, Juan G. Oliveira, Horácio A. B. F. Souto, Eduardo |
author_sort | Bragança, Hendrio |
collection | PubMed |
description | In this article, we introduce explainable methods to understand how Human Activity Recognition (HAR) mobile systems perform based on the chosen validation strategies. Our results introduce a new way to discover potential bias problems that overestimate the prediction accuracy of an algorithm because of the inappropriate choice of validation methodology. We show how the SHAP (Shapley additive explanations) framework, used in literature to explain the predictions of any machine learning model, presents itself as a tool that can provide graphical insights into how human activity recognition models achieve their results. Now it is possible to analyze which features are important to a HAR system in each validation methodology in a simplified way. We not only demonstrate that the validation procedure k-folds cross-validation (k-CV), used in most works to evaluate the expected error in a HAR system, can overestimate by about 13% the prediction accuracy in three public datasets but also choose a different feature set when compared with the universal model. Combining explainable methods with machine learning algorithms has the potential to help new researchers look inside the decisions of the machine learning algorithms, avoiding most times the overestimation of prediction accuracy, understanding relations between features, and finding bias before deploying the system in real-world scenarios. |
format | Online Article Text |
id | pubmed-8954513 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-89545132022-03-26 How Validation Methodology Influences Human Activity Recognition Mobile Systems Bragança, Hendrio Colonna, Juan G. Oliveira, Horácio A. B. F. Souto, Eduardo Sensors (Basel) Article In this article, we introduce explainable methods to understand how Human Activity Recognition (HAR) mobile systems perform based on the chosen validation strategies. Our results introduce a new way to discover potential bias problems that overestimate the prediction accuracy of an algorithm because of the inappropriate choice of validation methodology. We show how the SHAP (Shapley additive explanations) framework, used in literature to explain the predictions of any machine learning model, presents itself as a tool that can provide graphical insights into how human activity recognition models achieve their results. Now it is possible to analyze which features are important to a HAR system in each validation methodology in a simplified way. We not only demonstrate that the validation procedure k-folds cross-validation (k-CV), used in most works to evaluate the expected error in a HAR system, can overestimate by about 13% the prediction accuracy in three public datasets but also choose a different feature set when compared with the universal model. Combining explainable methods with machine learning algorithms has the potential to help new researchers look inside the decisions of the machine learning algorithms, avoiding most times the overestimation of prediction accuracy, understanding relations between features, and finding bias before deploying the system in real-world scenarios. MDPI 2022-03-18 /pmc/articles/PMC8954513/ /pubmed/35336529 http://dx.doi.org/10.3390/s22062360 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Bragança, Hendrio Colonna, Juan G. Oliveira, Horácio A. B. F. Souto, Eduardo How Validation Methodology Influences Human Activity Recognition Mobile Systems |
title | How Validation Methodology Influences Human Activity Recognition Mobile Systems |
title_full | How Validation Methodology Influences Human Activity Recognition Mobile Systems |
title_fullStr | How Validation Methodology Influences Human Activity Recognition Mobile Systems |
title_full_unstemmed | How Validation Methodology Influences Human Activity Recognition Mobile Systems |
title_short | How Validation Methodology Influences Human Activity Recognition Mobile Systems |
title_sort | how validation methodology influences human activity recognition mobile systems |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8954513/ https://www.ncbi.nlm.nih.gov/pubmed/35336529 http://dx.doi.org/10.3390/s22062360 |
work_keys_str_mv | AT bragancahendrio howvalidationmethodologyinfluenceshumanactivityrecognitionmobilesystems AT colonnajuang howvalidationmethodologyinfluenceshumanactivityrecognitionmobilesystems AT oliveirahoracioabf howvalidationmethodologyinfluenceshumanactivityrecognitionmobilesystems AT soutoeduardo howvalidationmethodologyinfluenceshumanactivityrecognitionmobilesystems |