Cargando…

Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences

Human Action Recognition (HAR) is the classification of an action performed by a human. The goal of this study was to recognize human actions in action video sequences. We present a novel feature descriptor for HAR that involves multiple features and combining them using fusion technique. The major...

Descripción completa

Detalles Bibliográficos
Autores principales: Patel, Chirag I., Labana, Dileep, Pandya, Sharnil, Modi, Kirit, Ghayvat, Hemant, Awais, Muhammad
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7766717/
https://www.ncbi.nlm.nih.gov/pubmed/33353248
http://dx.doi.org/10.3390/s20247299
_version_ 1783628785565827072
author Patel, Chirag I.
Labana, Dileep
Pandya, Sharnil
Modi, Kirit
Ghayvat, Hemant
Awais, Muhammad
author_facet Patel, Chirag I.
Labana, Dileep
Pandya, Sharnil
Modi, Kirit
Ghayvat, Hemant
Awais, Muhammad
author_sort Patel, Chirag I.
collection PubMed
description Human Action Recognition (HAR) is the classification of an action performed by a human. The goal of this study was to recognize human actions in action video sequences. We present a novel feature descriptor for HAR that involves multiple features and combining them using fusion technique. The major focus of the feature descriptor is to exploits the action dissimilarities. The key contribution of the proposed approach is to built robust features descriptor that can work for underlying video sequences and various classification models. To achieve the objective of the proposed work, HAR has been performed in the following manner. First, moving object detection and segmentation are performed from the background. The features are calculated using the histogram of oriented gradient (HOG) from a segmented moving object. To reduce the feature descriptor size, we take an averaging of the HOG features across non-overlapping video frames. For the frequency domain information we have calculated regional features from the Fourier hog. Moreover, we have also included the velocity and displacement of moving object. Finally, we use fusion technique to combine these features in the proposed work. After a feature descriptor is prepared, it is provided to the classifier. Here, we have used well-known classifiers such as artificial neural networks (ANNs), support vector machine (SVM), multiple kernel learning (MKL), Meta-cognitive Neural Network (McNN), and the late fusion methods. The main objective of the proposed approach is to prepare a robust feature descriptor and to show the diversity of our feature descriptor. Though we are using five different classifiers, our feature descriptor performs relatively well across the various classifiers. The proposed approach is performed and compared with the state-of-the-art methods for action recognition on two publicly available benchmark datasets (KTH and Weizmann) and for cross-validation on the UCF11 dataset, HMDB51 dataset, and UCF101 dataset. Results of the control experiments, such as a change in the SVM classifier and the effects of the second hidden layer in ANN, are also reported. The results demonstrate that the proposed method performs reasonably compared with the majority of existing state-of-the-art methods, including the convolutional neural network-based feature extractors.
format Online
Article
Text
id pubmed-7766717
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-77667172020-12-28 Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences Patel, Chirag I. Labana, Dileep Pandya, Sharnil Modi, Kirit Ghayvat, Hemant Awais, Muhammad Sensors (Basel) Article Human Action Recognition (HAR) is the classification of an action performed by a human. The goal of this study was to recognize human actions in action video sequences. We present a novel feature descriptor for HAR that involves multiple features and combining them using fusion technique. The major focus of the feature descriptor is to exploits the action dissimilarities. The key contribution of the proposed approach is to built robust features descriptor that can work for underlying video sequences and various classification models. To achieve the objective of the proposed work, HAR has been performed in the following manner. First, moving object detection and segmentation are performed from the background. The features are calculated using the histogram of oriented gradient (HOG) from a segmented moving object. To reduce the feature descriptor size, we take an averaging of the HOG features across non-overlapping video frames. For the frequency domain information we have calculated regional features from the Fourier hog. Moreover, we have also included the velocity and displacement of moving object. Finally, we use fusion technique to combine these features in the proposed work. After a feature descriptor is prepared, it is provided to the classifier. Here, we have used well-known classifiers such as artificial neural networks (ANNs), support vector machine (SVM), multiple kernel learning (MKL), Meta-cognitive Neural Network (McNN), and the late fusion methods. The main objective of the proposed approach is to prepare a robust feature descriptor and to show the diversity of our feature descriptor. Though we are using five different classifiers, our feature descriptor performs relatively well across the various classifiers. The proposed approach is performed and compared with the state-of-the-art methods for action recognition on two publicly available benchmark datasets (KTH and Weizmann) and for cross-validation on the UCF11 dataset, HMDB51 dataset, and UCF101 dataset. Results of the control experiments, such as a change in the SVM classifier and the effects of the second hidden layer in ANN, are also reported. The results demonstrate that the proposed method performs reasonably compared with the majority of existing state-of-the-art methods, including the convolutional neural network-based feature extractors. MDPI 2020-12-18 /pmc/articles/PMC7766717/ /pubmed/33353248 http://dx.doi.org/10.3390/s20247299 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Patel, Chirag I.
Labana, Dileep
Pandya, Sharnil
Modi, Kirit
Ghayvat, Hemant
Awais, Muhammad
Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences
title Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences
title_full Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences
title_fullStr Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences
title_full_unstemmed Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences
title_short Histogram of Oriented Gradient-Based Fusion of Features for Human Action Recognition in Action Video Sequences
title_sort histogram of oriented gradient-based fusion of features for human action recognition in action video sequences
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7766717/
https://www.ncbi.nlm.nih.gov/pubmed/33353248
http://dx.doi.org/10.3390/s20247299
work_keys_str_mv AT patelchiragi histogramoforientedgradientbasedfusionoffeaturesforhumanactionrecognitioninactionvideosequences
AT labanadileep histogramoforientedgradientbasedfusionoffeaturesforhumanactionrecognitioninactionvideosequences
AT pandyasharnil histogramoforientedgradientbasedfusionoffeaturesforhumanactionrecognitioninactionvideosequences
AT modikirit histogramoforientedgradientbasedfusionoffeaturesforhumanactionrecognitioninactionvideosequences
AT ghayvathemant histogramoforientedgradientbasedfusionoffeaturesforhumanactionrecognitioninactionvideosequences
AT awaismuhammad histogramoforientedgradientbasedfusionoffeaturesforhumanactionrecognitioninactionvideosequences