Cargando…
Attention-Based Temporal Encoding Network with Background-Independent Motion Mask for Action Recognition
Convolutional neural network (CNN) has been leaping forward in recent years. However, the high dimensionality, rich human dynamic characteristics, and various kinds of background interference increase difficulty for traditional CNNs in capturing complicated motion data in videos. A novel framework n...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8024088/ https://www.ncbi.nlm.nih.gov/pubmed/33859682 http://dx.doi.org/10.1155/2021/8890808 |
_version_ | 1783675239203340288 |
---|---|
author | Weng, Zhengkui Jin, Zhipeng Chen, Shuangxi Shen, Quanquan Ren, Xiangyang Li, Wuzhao |
author_facet | Weng, Zhengkui Jin, Zhipeng Chen, Shuangxi Shen, Quanquan Ren, Xiangyang Li, Wuzhao |
author_sort | Weng, Zhengkui |
collection | PubMed |
description | Convolutional neural network (CNN) has been leaping forward in recent years. However, the high dimensionality, rich human dynamic characteristics, and various kinds of background interference increase difficulty for traditional CNNs in capturing complicated motion data in videos. A novel framework named the attention-based temporal encoding network (ATEN) with background-independent motion mask (BIMM) is proposed to achieve video action recognition here. Initially, we introduce one motion segmenting approach on the basis of boundary prior by associating with the minimal geodesic distance inside a weighted graph that is not directed. Then, we propose one dynamic contrast segmenting strategic procedure for segmenting the object that moves within complicated environments. Subsequently, we build the BIMM for enhancing the object that moves based on the suppression of the not relevant background inside the respective frame. Furthermore, we design one long-range attention system inside ATEN, capable of effectively remedying the dependency of sophisticated actions that are not periodic in a long term based on the more automatic focus on the semantical vital frames other than the equal process for overall sampled frames. For this reason, the attention mechanism is capable of suppressing the temporal redundancy and highlighting the discriminative frames. Lastly, the framework is assessed by using HMDB51 and UCF101 datasets. As revealed from the experimentally achieved results, our ATEN with BIMM gains 94.5% and 70.6% accuracy, respectively, which outperforms a number of existing methods on both datasets. |
format | Online Article Text |
id | pubmed-8024088 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-80240882021-04-14 Attention-Based Temporal Encoding Network with Background-Independent Motion Mask for Action Recognition Weng, Zhengkui Jin, Zhipeng Chen, Shuangxi Shen, Quanquan Ren, Xiangyang Li, Wuzhao Comput Intell Neurosci Research Article Convolutional neural network (CNN) has been leaping forward in recent years. However, the high dimensionality, rich human dynamic characteristics, and various kinds of background interference increase difficulty for traditional CNNs in capturing complicated motion data in videos. A novel framework named the attention-based temporal encoding network (ATEN) with background-independent motion mask (BIMM) is proposed to achieve video action recognition here. Initially, we introduce one motion segmenting approach on the basis of boundary prior by associating with the minimal geodesic distance inside a weighted graph that is not directed. Then, we propose one dynamic contrast segmenting strategic procedure for segmenting the object that moves within complicated environments. Subsequently, we build the BIMM for enhancing the object that moves based on the suppression of the not relevant background inside the respective frame. Furthermore, we design one long-range attention system inside ATEN, capable of effectively remedying the dependency of sophisticated actions that are not periodic in a long term based on the more automatic focus on the semantical vital frames other than the equal process for overall sampled frames. For this reason, the attention mechanism is capable of suppressing the temporal redundancy and highlighting the discriminative frames. Lastly, the framework is assessed by using HMDB51 and UCF101 datasets. As revealed from the experimentally achieved results, our ATEN with BIMM gains 94.5% and 70.6% accuracy, respectively, which outperforms a number of existing methods on both datasets. Hindawi 2021-03-27 /pmc/articles/PMC8024088/ /pubmed/33859682 http://dx.doi.org/10.1155/2021/8890808 Text en Copyright © 2021 Zhengkui Weng et al. https://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Weng, Zhengkui Jin, Zhipeng Chen, Shuangxi Shen, Quanquan Ren, Xiangyang Li, Wuzhao Attention-Based Temporal Encoding Network with Background-Independent Motion Mask for Action Recognition |
title | Attention-Based Temporal Encoding Network with Background-Independent Motion Mask for Action Recognition |
title_full | Attention-Based Temporal Encoding Network with Background-Independent Motion Mask for Action Recognition |
title_fullStr | Attention-Based Temporal Encoding Network with Background-Independent Motion Mask for Action Recognition |
title_full_unstemmed | Attention-Based Temporal Encoding Network with Background-Independent Motion Mask for Action Recognition |
title_short | Attention-Based Temporal Encoding Network with Background-Independent Motion Mask for Action Recognition |
title_sort | attention-based temporal encoding network with background-independent motion mask for action recognition |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8024088/ https://www.ncbi.nlm.nih.gov/pubmed/33859682 http://dx.doi.org/10.1155/2021/8890808 |
work_keys_str_mv | AT wengzhengkui attentionbasedtemporalencodingnetworkwithbackgroundindependentmotionmaskforactionrecognition AT jinzhipeng attentionbasedtemporalencodingnetworkwithbackgroundindependentmotionmaskforactionrecognition AT chenshuangxi attentionbasedtemporalencodingnetworkwithbackgroundindependentmotionmaskforactionrecognition AT shenquanquan attentionbasedtemporalencodingnetworkwithbackgroundindependentmotionmaskforactionrecognition AT renxiangyang attentionbasedtemporalencodingnetworkwithbackgroundindependentmotionmaskforactionrecognition AT liwuzhao attentionbasedtemporalencodingnetworkwithbackgroundindependentmotionmaskforactionrecognition |