Cargando…
Keys for Action: An Efficient Keyframe-Based Approach for 3D Action Recognition Using a Deep Neural Network
In this paper, we propose a novel and efficient framework for 3D action recognition using a deep learning architecture. First, we develop a 3D normalized pose space that consists of only 3D normalized poses, which are generated by discarding translation and orientation information. From these poses,...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7218879/ https://www.ncbi.nlm.nih.gov/pubmed/32326468 http://dx.doi.org/10.3390/s20082226 |
_version_ | 1783532882296307712 |
---|---|
author | Yasin, Hashim Hussain, Mazhar Weber, Andreas |
author_facet | Yasin, Hashim Hussain, Mazhar Weber, Andreas |
author_sort | Yasin, Hashim |
collection | PubMed |
description | In this paper, we propose a novel and efficient framework for 3D action recognition using a deep learning architecture. First, we develop a 3D normalized pose space that consists of only 3D normalized poses, which are generated by discarding translation and orientation information. From these poses, we extract joint features and employ them further in a Deep Neural Network (DNN) in order to learn the action model. The architecture of our DNN consists of two hidden layers with the sigmoid activation function and an output layer with the softmax function. Furthermore, we propose a keyframe extraction methodology through which, from a motion sequence of 3D frames, we efficiently extract the keyframes that contribute substantially to the performance of the action. In this way, we eliminate redundant frames and reduce the length of the motion. More precisely, we ultimately summarize the motion sequence, while preserving the original motion semantics. We only consider the remaining essential informative frames in the process of action recognition, and the proposed pipeline is sufficiently fast and robust as a result. Finally, we evaluate our proposed framework intensively on publicly available benchmark Motion Capture (MoCap) datasets, namely HDM05 and CMU. From our experiments, we reveal that our proposed scheme significantly outperforms other state-of-the-art approaches. |
format | Online Article Text |
id | pubmed-7218879 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-72188792020-05-22 Keys for Action: An Efficient Keyframe-Based Approach for 3D Action Recognition Using a Deep Neural Network Yasin, Hashim Hussain, Mazhar Weber, Andreas Sensors (Basel) Article In this paper, we propose a novel and efficient framework for 3D action recognition using a deep learning architecture. First, we develop a 3D normalized pose space that consists of only 3D normalized poses, which are generated by discarding translation and orientation information. From these poses, we extract joint features and employ them further in a Deep Neural Network (DNN) in order to learn the action model. The architecture of our DNN consists of two hidden layers with the sigmoid activation function and an output layer with the softmax function. Furthermore, we propose a keyframe extraction methodology through which, from a motion sequence of 3D frames, we efficiently extract the keyframes that contribute substantially to the performance of the action. In this way, we eliminate redundant frames and reduce the length of the motion. More precisely, we ultimately summarize the motion sequence, while preserving the original motion semantics. We only consider the remaining essential informative frames in the process of action recognition, and the proposed pipeline is sufficiently fast and robust as a result. Finally, we evaluate our proposed framework intensively on publicly available benchmark Motion Capture (MoCap) datasets, namely HDM05 and CMU. From our experiments, we reveal that our proposed scheme significantly outperforms other state-of-the-art approaches. MDPI 2020-04-15 /pmc/articles/PMC7218879/ /pubmed/32326468 http://dx.doi.org/10.3390/s20082226 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Yasin, Hashim Hussain, Mazhar Weber, Andreas Keys for Action: An Efficient Keyframe-Based Approach for 3D Action Recognition Using a Deep Neural Network |
title | Keys for Action: An Efficient Keyframe-Based Approach for 3D Action Recognition Using a Deep Neural Network |
title_full | Keys for Action: An Efficient Keyframe-Based Approach for 3D Action Recognition Using a Deep Neural Network |
title_fullStr | Keys for Action: An Efficient Keyframe-Based Approach for 3D Action Recognition Using a Deep Neural Network |
title_full_unstemmed | Keys for Action: An Efficient Keyframe-Based Approach for 3D Action Recognition Using a Deep Neural Network |
title_short | Keys for Action: An Efficient Keyframe-Based Approach for 3D Action Recognition Using a Deep Neural Network |
title_sort | keys for action: an efficient keyframe-based approach for 3d action recognition using a deep neural network |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7218879/ https://www.ncbi.nlm.nih.gov/pubmed/32326468 http://dx.doi.org/10.3390/s20082226 |
work_keys_str_mv | AT yasinhashim keysforactionanefficientkeyframebasedapproachfor3dactionrecognitionusingadeepneuralnetwork AT hussainmazhar keysforactionanefficientkeyframebasedapproachfor3dactionrecognitionusingadeepneuralnetwork AT weberandreas keysforactionanefficientkeyframebasedapproachfor3dactionrecognitionusingadeepneuralnetwork |