Cargando…
New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition †
For the effective application of thriving human-assistive technologies in healthcare services and human–robot collaborative tasks, computing devices must be aware of human movements. Developing a reliable real-time activity recognition method for the continuous and smooth operation of such smart dev...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8073736/ https://www.ncbi.nlm.nih.gov/pubmed/33923706 http://dx.doi.org/10.3390/s21082814 |
_version_ | 1783684198789283840 |
---|---|
author | Alemayoh, Tsige Tadesse Lee, Jae Hoon Okamoto, Shingo |
author_facet | Alemayoh, Tsige Tadesse Lee, Jae Hoon Okamoto, Shingo |
author_sort | Alemayoh, Tsige Tadesse |
collection | PubMed |
description | For the effective application of thriving human-assistive technologies in healthcare services and human–robot collaborative tasks, computing devices must be aware of human movements. Developing a reliable real-time activity recognition method for the continuous and smooth operation of such smart devices is imperative. To achieve this, light and intelligent methods that use ubiquitous sensors are pivotal. In this study, with the correlation of time series data in mind, a new method of data structuring for deeper feature extraction is introduced herein. The activity data were collected using a smartphone with the help of an exclusively developed iOS application. Data from eight activities were shaped into single and double-channels to extract deep temporal and spatial features of the signals. In addition to the time domain, raw data were represented via the Fourier and wavelet domains. Among the several neural network models used to fit the deep-learning classification of the activities, a convolutional neural network with a double-channeled time-domain input performed well. This method was further evaluated using other public datasets, and better performance was obtained. The practicability of the trained model was finally tested on a computer and a smartphone in real-time, where it demonstrated promising results. |
format | Online Article Text |
id | pubmed-8073736 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-80737362021-04-27 New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition † Alemayoh, Tsige Tadesse Lee, Jae Hoon Okamoto, Shingo Sensors (Basel) Article For the effective application of thriving human-assistive technologies in healthcare services and human–robot collaborative tasks, computing devices must be aware of human movements. Developing a reliable real-time activity recognition method for the continuous and smooth operation of such smart devices is imperative. To achieve this, light and intelligent methods that use ubiquitous sensors are pivotal. In this study, with the correlation of time series data in mind, a new method of data structuring for deeper feature extraction is introduced herein. The activity data were collected using a smartphone with the help of an exclusively developed iOS application. Data from eight activities were shaped into single and double-channels to extract deep temporal and spatial features of the signals. In addition to the time domain, raw data were represented via the Fourier and wavelet domains. Among the several neural network models used to fit the deep-learning classification of the activities, a convolutional neural network with a double-channeled time-domain input performed well. This method was further evaluated using other public datasets, and better performance was obtained. The practicability of the trained model was finally tested on a computer and a smartphone in real-time, where it demonstrated promising results. MDPI 2021-04-16 /pmc/articles/PMC8073736/ /pubmed/33923706 http://dx.doi.org/10.3390/s21082814 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Alemayoh, Tsige Tadesse Lee, Jae Hoon Okamoto, Shingo New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition † |
title | New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition † |
title_full | New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition † |
title_fullStr | New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition † |
title_full_unstemmed | New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition † |
title_short | New Sensor Data Structuring for Deeper Feature Extraction in Human Activity Recognition † |
title_sort | new sensor data structuring for deeper feature extraction in human activity recognition † |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8073736/ https://www.ncbi.nlm.nih.gov/pubmed/33923706 http://dx.doi.org/10.3390/s21082814 |
work_keys_str_mv | AT alemayohtsigetadesse newsensordatastructuringfordeeperfeatureextractioninhumanactivityrecognition AT leejaehoon newsensordatastructuringfordeeperfeatureextractioninhumanactivityrecognition AT okamotoshingo newsensordatastructuringfordeeperfeatureextractioninhumanactivityrecognition |