Cargando…

New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition †

For the effective application of thriving human-assistive technologies in healthcare services and human–robot collaborative tasks, computing devices must be aware of human movements. Developing a reliable real-time activity recognition method for the continuous and smooth operation of such smart dev...

Descripción completa

Detalles Bibliográficos
Autores principales: Alemayoh, Tsige Tadesse, Lee, Jae Hoon, Okamoto, Shingo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8073736/
https://www.ncbi.nlm.nih.gov/pubmed/33923706
http://dx.doi.org/10.3390/s21082814
_version_ 1783684198789283840
author Alemayoh, Tsige Tadesse
Lee, Jae Hoon
Okamoto, Shingo
author_facet Alemayoh, Tsige Tadesse
Lee, Jae Hoon
Okamoto, Shingo
author_sort Alemayoh, Tsige Tadesse
collection PubMed
description For the effective application of thriving human-assistive technologies in healthcare services and human–robot collaborative tasks, computing devices must be aware of human movements. Developing a reliable real-time activity recognition method for the continuous and smooth operation of such smart devices is imperative. To achieve this, light and intelligent methods that use ubiquitous sensors are pivotal. In this study, with the correlation of time series data in mind, a new method of data structuring for deeper feature extraction is introduced herein. The activity data were collected using a smartphone with the help of an exclusively developed iOS application. Data from eight activities were shaped into single and double-channels to extract deep temporal and spatial features of the signals. In addition to the time domain, raw data were represented via the Fourier and wavelet domains. Among the several neural network models used to fit the deep-learning classification of the activities, a convolutional neural network with a double-channeled time-domain input performed well. This method was further evaluated using other public datasets, and better performance was obtained. The practicability of the trained model was finally tested on a computer and a smartphone in real-time, where it demonstrated promising results.
format Online
Article
Text
id pubmed-8073736
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-80737362021-04-27 New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition † Alemayoh, Tsige Tadesse Lee, Jae Hoon Okamoto, Shingo Sensors (Basel) Article For the effective application of thriving human-assistive technologies in healthcare services and human–robot collaborative tasks, computing devices must be aware of human movements. Developing a reliable real-time activity recognition method for the continuous and smooth operation of such smart devices is imperative. To achieve this, light and intelligent methods that use ubiquitous sensors are pivotal. In this study, with the correlation of time series data in mind, a new method of data structuring for deeper feature extraction is introduced herein. The activity data were collected using a smartphone with the help of an exclusively developed iOS application. Data from eight activities were shaped into single and double-channels to extract deep temporal and spatial features of the signals. In addition to the time domain, raw data were represented via the Fourier and wavelet domains. Among the several neural network models used to fit the deep-learning classification of the activities, a convolutional neural network with a double-channeled time-domain input performed well. This method was further evaluated using other public datasets, and better performance was obtained. The practicability of the trained model was finally tested on a computer and a smartphone in real-time, where it demonstrated promising results. MDPI 2021-04-16 /pmc/articles/PMC8073736/ /pubmed/33923706 http://dx.doi.org/10.3390/s21082814 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Alemayoh, Tsige Tadesse
Lee, Jae Hoon
Okamoto, Shingo
New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition †
title New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition †
title_full New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition †
title_fullStr New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition †
title_full_unstemmed New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition †
title_short New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition †
title_sort new sensor data structuring for deeper feature extraction in human activity recognition †
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8073736/
https://www.ncbi.nlm.nih.gov/pubmed/33923706
http://dx.doi.org/10.3390/s21082814
work_keys_str_mv AT alemayohtsigetadesse newsensordatastructuringfordeeperfeatureextractioninhumanactivityrecognition
AT leejaehoon newsensordatastructuringfordeeperfeatureextractioninhumanactivityrecognition
AT okamotoshingo newsensordatastructuringfordeeperfeatureextractioninhumanactivityrecognition