Cargando…
KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition
Human action recognition is an essential process in surveillance video analysis, which is used to understand the behavior of people to ensure safety. Most of the existing methods for HAR use computationally heavy networks such as 3D CNN and two-stream networks. To alleviate the challenges in the imp...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10297237/ https://www.ncbi.nlm.nih.gov/pubmed/37372188 http://dx.doi.org/10.3390/e25060844 |
_version_ | 1785063835908964352 |
---|---|
author | Paramasivam, Kalaivani Sindha, Mohamed Mansoor Roomi Balakrishnan, Sathya Bama |
author_facet | Paramasivam, Kalaivani Sindha, Mohamed Mansoor Roomi Balakrishnan, Sathya Bama |
author_sort | Paramasivam, Kalaivani |
collection | PubMed |
description | Human action recognition is an essential process in surveillance video analysis, which is used to understand the behavior of people to ensure safety. Most of the existing methods for HAR use computationally heavy networks such as 3D CNN and two-stream networks. To alleviate the challenges in the implementation and training of 3D deep learning networks, which have more parameters, a customized lightweight directed acyclic graph-based residual 2D CNN with fewer parameters was designed from scratch and named HARNet. A novel pipeline for the construction of spatial motion data from raw video input is presented for the latent representation learning of human actions. The constructed input is fed to the network for simultaneous operation over spatial and motion information in a single stream, and the latent representation learned at the fully connected layer is extracted and fed to the conventional machine learning classifiers for action recognition. The proposed work was empirically verified, and the experimental results were compared with those for existing methods. The results show that the proposed method outperforms state-of-the-art (SOTA) methods with a percentage improvement of 2.75% on UCF101, 10.94% on HMDB51, and 0.18% on the KTH dataset. |
format | Online Article Text |
id | pubmed-10297237 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-102972372023-06-28 KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition Paramasivam, Kalaivani Sindha, Mohamed Mansoor Roomi Balakrishnan, Sathya Bama Entropy (Basel) Article Human action recognition is an essential process in surveillance video analysis, which is used to understand the behavior of people to ensure safety. Most of the existing methods for HAR use computationally heavy networks such as 3D CNN and two-stream networks. To alleviate the challenges in the implementation and training of 3D deep learning networks, which have more parameters, a customized lightweight directed acyclic graph-based residual 2D CNN with fewer parameters was designed from scratch and named HARNet. A novel pipeline for the construction of spatial motion data from raw video input is presented for the latent representation learning of human actions. The constructed input is fed to the network for simultaneous operation over spatial and motion information in a single stream, and the latent representation learned at the fully connected layer is extracted and fed to the conventional machine learning classifiers for action recognition. The proposed work was empirically verified, and the experimental results were compared with those for existing methods. The results show that the proposed method outperforms state-of-the-art (SOTA) methods with a percentage improvement of 2.75% on UCF101, 10.94% on HMDB51, and 0.18% on the KTH dataset. MDPI 2023-05-25 /pmc/articles/PMC10297237/ /pubmed/37372188 http://dx.doi.org/10.3390/e25060844 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Paramasivam, Kalaivani Sindha, Mohamed Mansoor Roomi Balakrishnan, Sathya Bama KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition |
title | KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition |
title_full | KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition |
title_fullStr | KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition |
title_full_unstemmed | KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition |
title_short | KNN-Based Machine Learning Classifier Used on Deep Learned Spatial Motion Features for Human Action Recognition |
title_sort | knn-based machine learning classifier used on deep learned spatial motion features for human action recognition |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10297237/ https://www.ncbi.nlm.nih.gov/pubmed/37372188 http://dx.doi.org/10.3390/e25060844 |
work_keys_str_mv | AT paramasivamkalaivani knnbasedmachinelearningclassifierusedondeeplearnedspatialmotionfeaturesforhumanactionrecognition AT sindhamohamedmansoorroomi knnbasedmachinelearningclassifierusedondeeplearnedspatialmotionfeaturesforhumanactionrecognition AT balakrishnansathyabama knnbasedmachinelearningclassifierusedondeeplearnedspatialmotionfeaturesforhumanactionrecognition |