Cargando…

Object Tracking in RGB-T Videos Using Modal-Aware Attention Network and Competitive Learning

Object tracking in RGB-thermal (RGB-T) videos is increasingly used in many fields due to the all-weather and all-day working capability of the dual-modality imaging system, as well as the rapid development of low-cost and miniaturized infrared camera technology. However, it is still very challenging...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, Hui, Zhang, Lei, Zhuo, Li, Zhang, Jing
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7014199/
https://www.ncbi.nlm.nih.gov/pubmed/32284517
http://dx.doi.org/10.3390/s20020393
_version_ 1783496574660247552
author Zhang, Hui
Zhang, Lei
Zhuo, Li
Zhang, Jing
author_facet Zhang, Hui
Zhang, Lei
Zhuo, Li
Zhang, Jing
author_sort Zhang, Hui
collection PubMed
description Object tracking in RGB-thermal (RGB-T) videos is increasingly used in many fields due to the all-weather and all-day working capability of the dual-modality imaging system, as well as the rapid development of low-cost and miniaturized infrared camera technology. However, it is still very challenging to effectively fuse dual-modality information to build a robust RGB-T tracker. In this paper, an RGB-T object tracking algorithm based on a modal-aware attention network and competitive learning (MaCNet) is proposed, which includes a feature extraction network, modal-aware attention network, and classification network. The feature extraction network adopts the form of a two-stream network to extract features from each modality image. The modal-aware attention network integrates the original data, establishes an attention model that characterizes the importance of different feature layers, and then guides the feature fusion to enhance the information interaction between modalities. The classification network constructs a modality-egoistic loss function through three parallel binary classifiers acting on the RGB branch, the thermal infrared branch, and the fusion branch, respectively. Guided by the training strategy of competitive learning, the entire network is fine-tuned in the direction of the optimal fusion of the dual modalities. Extensive experiments on several publicly available RGB-T datasets show that our tracker has superior performance compared to other latest RGB-T and RGB tracking approaches.
format Online
Article
Text
id pubmed-7014199
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-70141992020-03-09 Object Tracking in RGB-T Videos Using Modal-Aware Attention Network and Competitive Learning Zhang, Hui Zhang, Lei Zhuo, Li Zhang, Jing Sensors (Basel) Article Object tracking in RGB-thermal (RGB-T) videos is increasingly used in many fields due to the all-weather and all-day working capability of the dual-modality imaging system, as well as the rapid development of low-cost and miniaturized infrared camera technology. However, it is still very challenging to effectively fuse dual-modality information to build a robust RGB-T tracker. In this paper, an RGB-T object tracking algorithm based on a modal-aware attention network and competitive learning (MaCNet) is proposed, which includes a feature extraction network, modal-aware attention network, and classification network. The feature extraction network adopts the form of a two-stream network to extract features from each modality image. The modal-aware attention network integrates the original data, establishes an attention model that characterizes the importance of different feature layers, and then guides the feature fusion to enhance the information interaction between modalities. The classification network constructs a modality-egoistic loss function through three parallel binary classifiers acting on the RGB branch, the thermal infrared branch, and the fusion branch, respectively. Guided by the training strategy of competitive learning, the entire network is fine-tuned in the direction of the optimal fusion of the dual modalities. Extensive experiments on several publicly available RGB-T datasets show that our tracker has superior performance compared to other latest RGB-T and RGB tracking approaches. MDPI 2020-01-10 /pmc/articles/PMC7014199/ /pubmed/32284517 http://dx.doi.org/10.3390/s20020393 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhang, Hui
Zhang, Lei
Zhuo, Li
Zhang, Jing
Object Tracking in RGB-T Videos Using Modal-Aware Attention Network and Competitive Learning
title Object Tracking in RGB-T Videos Using Modal-Aware Attention Network and Competitive Learning
title_full Object Tracking in RGB-T Videos Using Modal-Aware Attention Network and Competitive Learning
title_fullStr Object Tracking in RGB-T Videos Using Modal-Aware Attention Network and Competitive Learning
title_full_unstemmed Object Tracking in RGB-T Videos Using Modal-Aware Attention Network and Competitive Learning
title_short Object Tracking in RGB-T Videos Using Modal-Aware Attention Network and Competitive Learning
title_sort object tracking in rgb-t videos using modal-aware attention network and competitive learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7014199/
https://www.ncbi.nlm.nih.gov/pubmed/32284517
http://dx.doi.org/10.3390/s20020393
work_keys_str_mv AT zhanghui objecttrackinginrgbtvideosusingmodalawareattentionnetworkandcompetitivelearning
AT zhanglei objecttrackinginrgbtvideosusingmodalawareattentionnetworkandcompetitivelearning
AT zhuoli objecttrackinginrgbtvideosusingmodalawareattentionnetworkandcompetitivelearning
AT zhangjing objecttrackinginrgbtvideosusingmodalawareattentionnetworkandcompetitivelearning