Cargando…
Feature Fusion of a Deep-Learning Algorithm into Wearable Sensor Devices for Human Activity Recognition
This paper presents a wearable device, fitted on the waist of a participant that recognizes six activities of daily living (walking, walking upstairs, walking downstairs, sitting, standing, and laying) through a deep-learning algorithm, human activity recognition (HAR). The wearable device comprises...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8706653/ https://www.ncbi.nlm.nih.gov/pubmed/34960388 http://dx.doi.org/10.3390/s21248294 |
_version_ | 1784622245411291136 |
---|---|
author | Yen, Chih-Ta Liao, Jia-Xian Huang, Yi-Kai |
author_facet | Yen, Chih-Ta Liao, Jia-Xian Huang, Yi-Kai |
author_sort | Yen, Chih-Ta |
collection | PubMed |
description | This paper presents a wearable device, fitted on the waist of a participant that recognizes six activities of daily living (walking, walking upstairs, walking downstairs, sitting, standing, and laying) through a deep-learning algorithm, human activity recognition (HAR). The wearable device comprises a single-board computer (SBC) and six-axis sensors. The deep-learning algorithm employs three parallel convolutional neural networks for local feature extraction and for subsequent concatenation to establish feature fusion models of varying kernel size. By using kernels of different sizes, relevant local features of varying lengths were identified, thereby increasing the accuracy of human activity recognition. Regarding experimental data, the database of University of California, Irvine (UCI) and self-recorded data were used separately. The self-recorded data were obtained by having 21 participants wear the device on their waist and perform six common activities in the laboratory. These data were used to verify the proposed deep-learning algorithm on the performance of the wearable device. The accuracy of these six activities in the UCI dataset and in the self-recorded data were 97.49% and 96.27%, respectively. The accuracies in tenfold cross-validation were 99.56% and 97.46%, respectively. The experimental results have successfully verified the proposed convolutional neural network (CNN) architecture, which can be used in rehabilitation assessment for people unable to exercise vigorously. |
format | Online Article Text |
id | pubmed-8706653 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-87066532021-12-25 Feature Fusion of a Deep-Learning Algorithm into Wearable Sensor Devices for Human Activity Recognition Yen, Chih-Ta Liao, Jia-Xian Huang, Yi-Kai Sensors (Basel) Article This paper presents a wearable device, fitted on the waist of a participant that recognizes six activities of daily living (walking, walking upstairs, walking downstairs, sitting, standing, and laying) through a deep-learning algorithm, human activity recognition (HAR). The wearable device comprises a single-board computer (SBC) and six-axis sensors. The deep-learning algorithm employs three parallel convolutional neural networks for local feature extraction and for subsequent concatenation to establish feature fusion models of varying kernel size. By using kernels of different sizes, relevant local features of varying lengths were identified, thereby increasing the accuracy of human activity recognition. Regarding experimental data, the database of University of California, Irvine (UCI) and self-recorded data were used separately. The self-recorded data were obtained by having 21 participants wear the device on their waist and perform six common activities in the laboratory. These data were used to verify the proposed deep-learning algorithm on the performance of the wearable device. The accuracy of these six activities in the UCI dataset and in the self-recorded data were 97.49% and 96.27%, respectively. The accuracies in tenfold cross-validation were 99.56% and 97.46%, respectively. The experimental results have successfully verified the proposed convolutional neural network (CNN) architecture, which can be used in rehabilitation assessment for people unable to exercise vigorously. MDPI 2021-12-11 /pmc/articles/PMC8706653/ /pubmed/34960388 http://dx.doi.org/10.3390/s21248294 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Yen, Chih-Ta Liao, Jia-Xian Huang, Yi-Kai Feature Fusion of a Deep-Learning Algorithm into Wearable Sensor Devices for Human Activity Recognition |
title | Feature Fusion of a Deep-Learning Algorithm into Wearable Sensor Devices for Human Activity Recognition |
title_full | Feature Fusion of a Deep-Learning Algorithm into Wearable Sensor Devices for Human Activity Recognition |
title_fullStr | Feature Fusion of a Deep-Learning Algorithm into Wearable Sensor Devices for Human Activity Recognition |
title_full_unstemmed | Feature Fusion of a Deep-Learning Algorithm into Wearable Sensor Devices for Human Activity Recognition |
title_short | Feature Fusion of a Deep-Learning Algorithm into Wearable Sensor Devices for Human Activity Recognition |
title_sort | feature fusion of a deep-learning algorithm into wearable sensor devices for human activity recognition |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8706653/ https://www.ncbi.nlm.nih.gov/pubmed/34960388 http://dx.doi.org/10.3390/s21248294 |
work_keys_str_mv | AT yenchihta featurefusionofadeeplearningalgorithmintowearablesensordevicesforhumanactivityrecognition AT liaojiaxian featurefusionofadeeplearningalgorithmintowearablesensordevicesforhumanactivityrecognition AT huangyikai featurefusionofadeeplearningalgorithmintowearablesensordevicesforhumanactivityrecognition |