Cargando…

KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition

Human action recognition is an essential process in surveillance video analysis, which is used to understand the behavior of people to ensure safety. Most of the existing methods for HAR use computationally heavy networks such as 3D CNN and two-stream networks. To alleviate the challenges in the imp...

Descripción completa

Detalles Bibliográficos
Autores principales: Paramasivam, Kalaivani, Sindha, Mohamed Mansoor Roomi, Balakrishnan, Sathya Bama
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10297237/
https://www.ncbi.nlm.nih.gov/pubmed/37372188
http://dx.doi.org/10.3390/e25060844
_version_ 1785063835908964352
author Paramasivam, Kalaivani
Sindha, Mohamed Mansoor Roomi
Balakrishnan, Sathya Bama
author_facet Paramasivam, Kalaivani
Sindha, Mohamed Mansoor Roomi
Balakrishnan, Sathya Bama
author_sort Paramasivam, Kalaivani
collection PubMed
description Human action recognition is an essential process in surveillance video analysis, which is used to understand the behavior of people to ensure safety. Most of the existing methods for HAR use computationally heavy networks such as 3D CNN and two-stream networks. To alleviate the challenges in the implementation and training of 3D deep learning networks, which have more parameters, a customized lightweight directed acyclic graph-based residual 2D CNN with fewer parameters was designed from scratch and named HARNet. A novel pipeline for the construction of spatial motion data from raw video input is presented for the latent representation learning of human actions. The constructed input is fed to the network for simultaneous operation over spatial and motion information in a single stream, and the latent representation learned at the fully connected layer is extracted and fed to the conventional machine learning classifiers for action recognition. The proposed work was empirically verified, and the experimental results were compared with those for existing methods. The results show that the proposed method outperforms state-of-the-art (SOTA) methods with a percentage improvement of 2.75% on UCF101, 10.94% on HMDB51, and 0.18% on the KTH dataset.
format Online
Article
Text
id pubmed-10297237
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-102972372023-06-28 KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition Paramasivam, Kalaivani Sindha, Mohamed Mansoor Roomi Balakrishnan, Sathya Bama Entropy (Basel) Article Human action recognition is an essential process in surveillance video analysis, which is used to understand the behavior of people to ensure safety. Most of the existing methods for HAR use computationally heavy networks such as 3D CNN and two-stream networks. To alleviate the challenges in the implementation and training of 3D deep learning networks, which have more parameters, a customized lightweight directed acyclic graph-based residual 2D CNN with fewer parameters was designed from scratch and named HARNet. A novel pipeline for the construction of spatial motion data from raw video input is presented for the latent representation learning of human actions. The constructed input is fed to the network for simultaneous operation over spatial and motion information in a single stream, and the latent representation learned at the fully connected layer is extracted and fed to the conventional machine learning classifiers for action recognition. The proposed work was empirically verified, and the experimental results were compared with those for existing methods. The results show that the proposed method outperforms state-of-the-art (SOTA) methods with a percentage improvement of 2.75% on UCF101, 10.94% on HMDB51, and 0.18% on the KTH dataset. MDPI 2023-05-25 /pmc/articles/PMC10297237/ /pubmed/37372188 http://dx.doi.org/10.3390/e25060844 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Paramasivam, Kalaivani
Sindha, Mohamed Mansoor Roomi
Balakrishnan, Sathya Bama
KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition
title KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition
title_full KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition
title_fullStr KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition
title_full_unstemmed KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition
title_short KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition
title_sort knn-based machine learning classifier used on deep learned spatial motion features for human action recognition
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10297237/
https://www.ncbi.nlm.nih.gov/pubmed/37372188
http://dx.doi.org/10.3390/e25060844
work_keys_str_mv AT paramasivamkalaivani knnbasedmachinelearningclassifierusedondeeplearnedspatialmotionfeaturesforhumanactionrecognition
AT sindhamohamedmansoorroomi knnbasedmachinelearningclassifierusedondeeplearnedspatialmotionfeaturesforhumanactionrecognition
AT balakrishnansathyabama knnbasedmachinelearningclassifierusedondeeplearnedspatialmotionfeaturesforhumanactionrecognition