Cargando…
Real-Time Human Activity Recognition with IMU and Encoder Sensors in Wearable Exoskeleton Robot via Deep Learning Networks
Wearable exoskeleton robots have become a promising technology for supporting human motions in multiple tasks. Activity recognition in real-time provides useful information to enhance the robot’s control assistance for daily tasks. This work implements a real-time activity recognition system based o...
Autores principales: | , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9783602/ https://www.ncbi.nlm.nih.gov/pubmed/36560059 http://dx.doi.org/10.3390/s22249690 |
_version_ | 1784857616293298176 |
---|---|
author | Jaramillo, Ismael Espinoza Jeong, Jin Gyun Lopez, Patricio Rivera Lee, Choong-Ho Kang, Do-Yeon Ha, Tae-Jun Oh, Ji-Heon Jung, Hwanseok Lee, Jin Hyuk Lee, Won Hee Kim, Tae-Seong |
author_facet | Jaramillo, Ismael Espinoza Jeong, Jin Gyun Lopez, Patricio Rivera Lee, Choong-Ho Kang, Do-Yeon Ha, Tae-Jun Oh, Ji-Heon Jung, Hwanseok Lee, Jin Hyuk Lee, Won Hee Kim, Tae-Seong |
author_sort | Jaramillo, Ismael Espinoza |
collection | PubMed |
description | Wearable exoskeleton robots have become a promising technology for supporting human motions in multiple tasks. Activity recognition in real-time provides useful information to enhance the robot’s control assistance for daily tasks. This work implements a real-time activity recognition system based on the activity signals of an inertial measurement unit (IMU) and a pair of rotary encoders integrated into the exoskeleton robot. Five deep learning models have been trained and evaluated for activity recognition. As a result, a subset of optimized deep learning models was transferred to an edge device for real-time evaluation in a continuous action environment using eight common human tasks: stand, bend, crouch, walk, sit-down, sit-up, and ascend and descend stairs. These eight robot wearer’s activities are recognized with an average accuracy of 97.35% in real-time tests, with an inference time under 10 ms and an overall latency of 0.506 s per recognition using the selected edge device. |
format | Online Article Text |
id | pubmed-9783602 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-97836022022-12-24 Real-Time Human Activity Recognition with IMU and Encoder Sensors in Wearable Exoskeleton Robot via Deep Learning Networks Jaramillo, Ismael Espinoza Jeong, Jin Gyun Lopez, Patricio Rivera Lee, Choong-Ho Kang, Do-Yeon Ha, Tae-Jun Oh, Ji-Heon Jung, Hwanseok Lee, Jin Hyuk Lee, Won Hee Kim, Tae-Seong Sensors (Basel) Article Wearable exoskeleton robots have become a promising technology for supporting human motions in multiple tasks. Activity recognition in real-time provides useful information to enhance the robot’s control assistance for daily tasks. This work implements a real-time activity recognition system based on the activity signals of an inertial measurement unit (IMU) and a pair of rotary encoders integrated into the exoskeleton robot. Five deep learning models have been trained and evaluated for activity recognition. As a result, a subset of optimized deep learning models was transferred to an edge device for real-time evaluation in a continuous action environment using eight common human tasks: stand, bend, crouch, walk, sit-down, sit-up, and ascend and descend stairs. These eight robot wearer’s activities are recognized with an average accuracy of 97.35% in real-time tests, with an inference time under 10 ms and an overall latency of 0.506 s per recognition using the selected edge device. MDPI 2022-12-10 /pmc/articles/PMC9783602/ /pubmed/36560059 http://dx.doi.org/10.3390/s22249690 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Jaramillo, Ismael Espinoza Jeong, Jin Gyun Lopez, Patricio Rivera Lee, Choong-Ho Kang, Do-Yeon Ha, Tae-Jun Oh, Ji-Heon Jung, Hwanseok Lee, Jin Hyuk Lee, Won Hee Kim, Tae-Seong Real-Time Human Activity Recognition with IMU and Encoder Sensors in Wearable Exoskeleton Robot via Deep Learning Networks |
title | Real-Time Human Activity Recognition with IMU and Encoder Sensors in Wearable Exoskeleton Robot via Deep Learning Networks |
title_full | Real-Time Human Activity Recognition with IMU and Encoder Sensors in Wearable Exoskeleton Robot via Deep Learning Networks |
title_fullStr | Real-Time Human Activity Recognition with IMU and Encoder Sensors in Wearable Exoskeleton Robot via Deep Learning Networks |
title_full_unstemmed | Real-Time Human Activity Recognition with IMU and Encoder Sensors in Wearable Exoskeleton Robot via Deep Learning Networks |
title_short | Real-Time Human Activity Recognition with IMU and Encoder Sensors in Wearable Exoskeleton Robot via Deep Learning Networks |
title_sort | real-time human activity recognition with imu and encoder sensors in wearable exoskeleton robot via deep learning networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9783602/ https://www.ncbi.nlm.nih.gov/pubmed/36560059 http://dx.doi.org/10.3390/s22249690 |
work_keys_str_mv | AT jaramilloismaelespinoza realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks AT jeongjingyun realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks AT lopezpatriciorivera realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks AT leechoongho realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks AT kangdoyeon realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks AT hataejun realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks AT ohjiheon realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks AT junghwanseok realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks AT leejinhyuk realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks AT leewonhee realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks AT kimtaeseong realtimehumanactivityrecognitionwithimuandencodersensorsinwearableexoskeletonrobotviadeeplearningnetworks |