Cargando…

Improving Wearable-Based Activity Recognition Using Image Representations †

Activity recognition based on inertial sensors is an essential task in mobile and ubiquitous computing. To date, the best performing approaches in this task are based on deep learning models. Although the performance of the approaches has been increasingly improving, a number of issues still remain....

Descripción completa

Detalles Bibliográficos
Autores principales: Sanchez Guinea, Alejandro, Sarabchian, Mehran, Mühlhäuser, Max
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8914937/
https://www.ncbi.nlm.nih.gov/pubmed/35270985
http://dx.doi.org/10.3390/s22051840
_version_ 1784667879802667008
author Sanchez Guinea, Alejandro
Sarabchian, Mehran
Mühlhäuser, Max
author_facet Sanchez Guinea, Alejandro
Sarabchian, Mehran
Mühlhäuser, Max
author_sort Sanchez Guinea, Alejandro
collection PubMed
description Activity recognition based on inertial sensors is an essential task in mobile and ubiquitous computing. To date, the best performing approaches in this task are based on deep learning models. Although the performance of the approaches has been increasingly improving, a number of issues still remain. Specifically, in this paper we focus on the issue of the dependence of today’s state-of-the-art approaches to complex ad hoc deep learning convolutional neural networks (CNNs), recurrent neural networks (RNNs), or a combination of both, which require specialized knowledge and considerable effort for their construction and optimal tuning. To address this issue, in this paper we propose an approach that automatically transforms the inertial sensors time-series data into images that represent in pixel form patterns found over time, allowing even a simple CNN to outperform complex ad hoc deep learning models that combine RNNs and CNNs for activity recognition. We conducted an extensive evaluation considering seven benchmark datasets that are among the most relevant in activity recognition. Our results demonstrate that our approach is able to outperform the state of the art in all cases, based on image representations that are generated through a process that is easy to implement, modify, and extend further, without the need of developing complex deep learning models.
format Online
Article
Text
id pubmed-8914937
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-89149372022-03-12 Improving Wearable-Based Activity Recognition Using Image Representations † Sanchez Guinea, Alejandro Sarabchian, Mehran Mühlhäuser, Max Sensors (Basel) Article Activity recognition based on inertial sensors is an essential task in mobile and ubiquitous computing. To date, the best performing approaches in this task are based on deep learning models. Although the performance of the approaches has been increasingly improving, a number of issues still remain. Specifically, in this paper we focus on the issue of the dependence of today’s state-of-the-art approaches to complex ad hoc deep learning convolutional neural networks (CNNs), recurrent neural networks (RNNs), or a combination of both, which require specialized knowledge and considerable effort for their construction and optimal tuning. To address this issue, in this paper we propose an approach that automatically transforms the inertial sensors time-series data into images that represent in pixel form patterns found over time, allowing even a simple CNN to outperform complex ad hoc deep learning models that combine RNNs and CNNs for activity recognition. We conducted an extensive evaluation considering seven benchmark datasets that are among the most relevant in activity recognition. Our results demonstrate that our approach is able to outperform the state of the art in all cases, based on image representations that are generated through a process that is easy to implement, modify, and extend further, without the need of developing complex deep learning models. MDPI 2022-02-25 /pmc/articles/PMC8914937/ /pubmed/35270985 http://dx.doi.org/10.3390/s22051840 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Sanchez Guinea, Alejandro
Sarabchian, Mehran
Mühlhäuser, Max
Improving Wearable-Based Activity Recognition Using Image Representations †
title Improving Wearable-Based Activity Recognition Using Image Representations †
title_full Improving Wearable-Based Activity Recognition Using Image Representations †
title_fullStr Improving Wearable-Based Activity Recognition Using Image Representations †
title_full_unstemmed Improving Wearable-Based Activity Recognition Using Image Representations †
title_short Improving Wearable-Based Activity Recognition Using Image Representations †
title_sort improving wearable-based activity recognition using image representations †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8914937/
https://www.ncbi.nlm.nih.gov/pubmed/35270985
http://dx.doi.org/10.3390/s22051840
work_keys_str_mv AT sanchezguineaalejandro improvingwearablebasedactivityrecognitionusingimagerepresentations
AT sarabchianmehran improvingwearablebasedactivityrecognitionusingimagerepresentations
AT muhlhausermax improvingwearablebasedactivityrecognitionusingimagerepresentations