Cargando…
Human action recognition method based on Motion Excitation and Temporal Aggregation module
Aiming at the problem of low modeling efficiency and feature loss of temporal modeling in human action recognition, we propose a human action recognition method based on Motion Excitation and Temporal Aggregation module (META). The method can capture multi-state and multi-scale temporal information...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9647446/ https://www.ncbi.nlm.nih.gov/pubmed/36387431 http://dx.doi.org/10.1016/j.heliyon.2022.e11401 |
_version_ | 1784827384461000704 |
---|---|
author | Ye, Qing Tan, Zexian Zhang, Yongmei |
author_facet | Ye, Qing Tan, Zexian Zhang, Yongmei |
author_sort | Ye, Qing |
collection | PubMed |
description | Aiming at the problem of low modeling efficiency and feature loss of temporal modeling in human action recognition, we propose a human action recognition method based on Motion Excitation and Temporal Aggregation module (META). The method can capture multi-state and multi-scale temporal information to achieve effective motion excitation. Firstly, temporal relational sampling is performed on video frames; Secondly, META is proposed to capture multi-state and multi-scale temporal information. META is composed of Multi-scale Motion Excitation module (MME) and Squeeze and Excitation Temporal Aggregation module (SETA). MME captures the feature level temporal difference by transforming the features into the temporal channel, which directly establishes the relationship between features and temporal channel, and solves the problem of low modeling efficiency. SETA transforms the local convolution into a set of sub-convolutions. Multiple sub-convolutions form hierarchies to extract features together and share the results of the upper convolutional layer, which increases the final temporal receptive field and solves the problem of feature loss. Moreover, the optical flow features are extracted through Cross modality pre-training to improve the utilization of temporal information. Finally, the result of human action recognition is carried out by combining spatiotemporal two stream features. Experimental results show that the accuracy of this method in UCF101 and HMDB-51 is 96.0% and 71.2% respectively, which is higher than other studies in the same period. |
format | Online Article Text |
id | pubmed-9647446 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Elsevier |
record_format | MEDLINE/PubMed |
spelling | pubmed-96474462022-11-15 Human action recognition method based on Motion Excitation and Temporal Aggregation module Ye, Qing Tan, Zexian Zhang, Yongmei Heliyon Research Article Aiming at the problem of low modeling efficiency and feature loss of temporal modeling in human action recognition, we propose a human action recognition method based on Motion Excitation and Temporal Aggregation module (META). The method can capture multi-state and multi-scale temporal information to achieve effective motion excitation. Firstly, temporal relational sampling is performed on video frames; Secondly, META is proposed to capture multi-state and multi-scale temporal information. META is composed of Multi-scale Motion Excitation module (MME) and Squeeze and Excitation Temporal Aggregation module (SETA). MME captures the feature level temporal difference by transforming the features into the temporal channel, which directly establishes the relationship between features and temporal channel, and solves the problem of low modeling efficiency. SETA transforms the local convolution into a set of sub-convolutions. Multiple sub-convolutions form hierarchies to extract features together and share the results of the upper convolutional layer, which increases the final temporal receptive field and solves the problem of feature loss. Moreover, the optical flow features are extracted through Cross modality pre-training to improve the utilization of temporal information. Finally, the result of human action recognition is carried out by combining spatiotemporal two stream features. Experimental results show that the accuracy of this method in UCF101 and HMDB-51 is 96.0% and 71.2% respectively, which is higher than other studies in the same period. Elsevier 2022-11-04 /pmc/articles/PMC9647446/ /pubmed/36387431 http://dx.doi.org/10.1016/j.heliyon.2022.e11401 Text en © 2022 The Author(s) https://creativecommons.org/licenses/by-nc-nd/4.0/This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). |
spellingShingle | Research Article Ye, Qing Tan, Zexian Zhang, Yongmei Human action recognition method based on Motion Excitation and Temporal Aggregation module |
title | Human action recognition method based on Motion Excitation and Temporal Aggregation module |
title_full | Human action recognition method based on Motion Excitation and Temporal Aggregation module |
title_fullStr | Human action recognition method based on Motion Excitation and Temporal Aggregation module |
title_full_unstemmed | Human action recognition method based on Motion Excitation and Temporal Aggregation module |
title_short | Human action recognition method based on Motion Excitation and Temporal Aggregation module |
title_sort | human action recognition method based on motion excitation and temporal aggregation module |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9647446/ https://www.ncbi.nlm.nih.gov/pubmed/36387431 http://dx.doi.org/10.1016/j.heliyon.2022.e11401 |
work_keys_str_mv | AT yeqing humanactionrecognitionmethodbasedonmotionexcitationandtemporalaggregationmodule AT tanzexian humanactionrecognitionmethodbasedonmotionexcitationandtemporalaggregationmodule AT zhangyongmei humanactionrecognitionmethodbasedonmotionexcitationandtemporalaggregationmodule |