Cargando…
Real-Time Human Action Recognition with a Low-Cost RGB Camera and Mobile Robot Platform
Human action recognition is an important research area in the field of computer vision that can be applied in surveillance, assisted living, and robotic systems interacting with people. Although various approaches have been widely used, recent studies have mainly focused on deep-learning networks us...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7287597/ https://www.ncbi.nlm.nih.gov/pubmed/32438776 http://dx.doi.org/10.3390/s20102886 |
_version_ | 1783545084924395520 |
---|---|
author | Lee, Junwoo Ahn, Bummo |
author_facet | Lee, Junwoo Ahn, Bummo |
author_sort | Lee, Junwoo |
collection | PubMed |
description | Human action recognition is an important research area in the field of computer vision that can be applied in surveillance, assisted living, and robotic systems interacting with people. Although various approaches have been widely used, recent studies have mainly focused on deep-learning networks using Kinect camera that can easily generate data on skeleton joints using depth data, and have achieved satisfactory performances. However, their models are deep and complex to achieve a higher recognition score; therefore, they cannot be applied to a mobile robot platform using a Kinect camera. To overcome these limitations, we suggest a method to classify human actions in real-time using a single RGB camera, which can be applied to the mobile robot platform as well. We integrated two open-source libraries, i.e., OpenPose and 3D-baseline, to extract skeleton joints on RGB images, and classified the actions using convolutional neural networks. Finally, we set up the mobile robot platform including an NVIDIA JETSON XAVIER embedded board and tracking algorithm to monitor a person continuously. We achieved an accuracy of 70% on the NTU-RGBD training dataset, and the whole process was performed on an average of 15 frames per second (FPS) on an embedded board system. |
format | Online Article Text |
id | pubmed-7287597 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-72875972020-06-15 Real-Time Human Action Recognition with a Low-Cost RGB Camera and Mobile Robot Platform Lee, Junwoo Ahn, Bummo Sensors (Basel) Article Human action recognition is an important research area in the field of computer vision that can be applied in surveillance, assisted living, and robotic systems interacting with people. Although various approaches have been widely used, recent studies have mainly focused on deep-learning networks using Kinect camera that can easily generate data on skeleton joints using depth data, and have achieved satisfactory performances. However, their models are deep and complex to achieve a higher recognition score; therefore, they cannot be applied to a mobile robot platform using a Kinect camera. To overcome these limitations, we suggest a method to classify human actions in real-time using a single RGB camera, which can be applied to the mobile robot platform as well. We integrated two open-source libraries, i.e., OpenPose and 3D-baseline, to extract skeleton joints on RGB images, and classified the actions using convolutional neural networks. Finally, we set up the mobile robot platform including an NVIDIA JETSON XAVIER embedded board and tracking algorithm to monitor a person continuously. We achieved an accuracy of 70% on the NTU-RGBD training dataset, and the whole process was performed on an average of 15 frames per second (FPS) on an embedded board system. MDPI 2020-05-19 /pmc/articles/PMC7287597/ /pubmed/32438776 http://dx.doi.org/10.3390/s20102886 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Lee, Junwoo Ahn, Bummo Real-Time Human Action Recognition with a Low-Cost RGB Camera and Mobile Robot Platform |
title | Real-Time Human Action Recognition with a Low-Cost RGB Camera and Mobile Robot Platform |
title_full | Real-Time Human Action Recognition with a Low-Cost RGB Camera and Mobile Robot Platform |
title_fullStr | Real-Time Human Action Recognition with a Low-Cost RGB Camera and Mobile Robot Platform |
title_full_unstemmed | Real-Time Human Action Recognition with a Low-Cost RGB Camera and Mobile Robot Platform |
title_short | Real-Time Human Action Recognition with a Low-Cost RGB Camera and Mobile Robot Platform |
title_sort | real-time human action recognition with a low-cost rgb camera and mobile robot platform |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7287597/ https://www.ncbi.nlm.nih.gov/pubmed/32438776 http://dx.doi.org/10.3390/s20102886 |
work_keys_str_mv | AT leejunwoo realtimehumanactionrecognitionwithalowcostrgbcameraandmobilerobotplatform AT ahnbummo realtimehumanactionrecognitionwithalowcostrgbcameraandmobilerobotplatform |