Cargando…

Deep Learning for Activity Recognition in Older People Using a Pocket-Worn Smartphone

Activity recognition can provide useful information about an older individual’s activity level and encourage older people to become more active to live longer in good health. This study aimed to develop an activity recognition algorithm for smartphone accelerometry data of older people. Deep learnin...

Descripción completa

Detalles Bibliográficos
Autores principales: Nan, Yashi, Lovell, Nigel H., Redmond, Stephen J., Wang, Kejia, Delbaere, Kim, van Schooten, Kimberley S.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7765519/
https://www.ncbi.nlm.nih.gov/pubmed/33334028
http://dx.doi.org/10.3390/s20247195
_version_ 1783628509123444736
author Nan, Yashi
Lovell, Nigel H.
Redmond, Stephen J.
Wang, Kejia
Delbaere, Kim
van Schooten, Kimberley S.
author_facet Nan, Yashi
Lovell, Nigel H.
Redmond, Stephen J.
Wang, Kejia
Delbaere, Kim
van Schooten, Kimberley S.
author_sort Nan, Yashi
collection PubMed
description Activity recognition can provide useful information about an older individual’s activity level and encourage older people to become more active to live longer in good health. This study aimed to develop an activity recognition algorithm for smartphone accelerometry data of older people. Deep learning algorithms, including convolutional neural network (CNN) and long short-term memory (LSTM), were evaluated in this study. Smartphone accelerometry data of free-living activities, performed by 53 older people (83.8 ± 3.8 years; 38 male) under standardized circumstances, were classified into lying, sitting, standing, transition, walking, walking upstairs, and walking downstairs. A 1D CNN, a multichannel CNN, a CNN-LSTM, and a multichannel CNN-LSTM model were tested. The models were compared on accuracy and computational efficiency. Results show that the multichannel CNN-LSTM model achieved the best classification results, with an 81.1% accuracy and an acceptable model and time complexity. Specifically, the accuracy was 67.0% for lying, 70.7% for sitting, 88.4% for standing, 78.2% for transitions, 88.7% for walking, 65.7% for walking downstairs, and 68.7% for walking upstairs. The findings indicated that the multichannel CNN-LSTM model was feasible for smartphone-based activity recognition in older people.
format Online
Article
Text
id pubmed-7765519
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-77655192020-12-27 Deep Learning for Activity Recognition in Older People Using a Pocket-Worn Smartphone Nan, Yashi Lovell, Nigel H. Redmond, Stephen J. Wang, Kejia Delbaere, Kim van Schooten, Kimberley S. Sensors (Basel) Article Activity recognition can provide useful information about an older individual’s activity level and encourage older people to become more active to live longer in good health. This study aimed to develop an activity recognition algorithm for smartphone accelerometry data of older people. Deep learning algorithms, including convolutional neural network (CNN) and long short-term memory (LSTM), were evaluated in this study. Smartphone accelerometry data of free-living activities, performed by 53 older people (83.8 ± 3.8 years; 38 male) under standardized circumstances, were classified into lying, sitting, standing, transition, walking, walking upstairs, and walking downstairs. A 1D CNN, a multichannel CNN, a CNN-LSTM, and a multichannel CNN-LSTM model were tested. The models were compared on accuracy and computational efficiency. Results show that the multichannel CNN-LSTM model achieved the best classification results, with an 81.1% accuracy and an acceptable model and time complexity. Specifically, the accuracy was 67.0% for lying, 70.7% for sitting, 88.4% for standing, 78.2% for transitions, 88.7% for walking, 65.7% for walking downstairs, and 68.7% for walking upstairs. The findings indicated that the multichannel CNN-LSTM model was feasible for smartphone-based activity recognition in older people. MDPI 2020-12-15 /pmc/articles/PMC7765519/ /pubmed/33334028 http://dx.doi.org/10.3390/s20247195 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Nan, Yashi
Lovell, Nigel H.
Redmond, Stephen J.
Wang, Kejia
Delbaere, Kim
van Schooten, Kimberley S.
Deep Learning for Activity Recognition in Older People Using a Pocket-Worn Smartphone
title Deep Learning for Activity Recognition in Older People Using a Pocket-Worn Smartphone
title_full Deep Learning for Activity Recognition in Older People Using a Pocket-Worn Smartphone
title_fullStr Deep Learning for Activity Recognition in Older People Using a Pocket-Worn Smartphone
title_full_unstemmed Deep Learning for Activity Recognition in Older People Using a Pocket-Worn Smartphone
title_short Deep Learning for Activity Recognition in Older People Using a Pocket-Worn Smartphone
title_sort deep learning for activity recognition in older people using a pocket-worn smartphone
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7765519/
https://www.ncbi.nlm.nih.gov/pubmed/33334028
http://dx.doi.org/10.3390/s20247195
work_keys_str_mv AT nanyashi deeplearningforactivityrecognitioninolderpeopleusingapocketwornsmartphone
AT lovellnigelh deeplearningforactivityrecognitioninolderpeopleusingapocketwornsmartphone
AT redmondstephenj deeplearningforactivityrecognitioninolderpeopleusingapocketwornsmartphone
AT wangkejia deeplearningforactivityrecognitioninolderpeopleusingapocketwornsmartphone
AT delbaerekim deeplearningforactivityrecognitioninolderpeopleusingapocketwornsmartphone
AT vanschootenkimberleys deeplearningforactivityrecognitioninolderpeopleusingapocketwornsmartphone