Cargando…

Siamese network with a depthwise over-parameterized convolutional layer for visual tracking

Visual tracking is a fundamental research task in vision computer. It has broad application prospects, such as military defense and civil security. Visual tracking encounters many challenges in practical application, such as occlusion, fast motion and background clutter. Siamese based trackers achie...

Descripción completa

Detalles Bibliográficos
Autores principales: Wang, Yuanyun, Zhang, Wenshuang, Zhang, Limin, Wang, Jun
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9432743/
https://www.ncbi.nlm.nih.gov/pubmed/36044439
http://dx.doi.org/10.1371/journal.pone.0273690
_version_ 1784780455004864512
author Wang, Yuanyun
Zhang, Wenshuang
Zhang, Limin
Wang, Jun
author_facet Wang, Yuanyun
Zhang, Wenshuang
Zhang, Limin
Wang, Jun
author_sort Wang, Yuanyun
collection PubMed
description Visual tracking is a fundamental research task in vision computer. It has broad application prospects, such as military defense and civil security. Visual tracking encounters many challenges in practical application, such as occlusion, fast motion and background clutter. Siamese based trackers achieve superior tracking performance in balanced accuracy and tracking speed. The deep feature extraction with Convolutional Neural Network (CNN) is an essential component in Siamese tracking framework. Although existing trackers take full advantage of deep feature information, the spatial structure and semantic information are not adequately exploited, which are helpful for enhancing target representations. The lack of these spatial and semantic information may lead to tracking drift. In this paper, we design a CNN feature extraction subnetwork based on a Depthwise Over-parameterized Convolutional layer (DO-Conv). A joint convolution method is introduced, namely the conventional and depthwise convolution. The depthwise convolution kernel explores independent channel information, which effectively extracts shallow spatial information and deep semantic information, and discards background information. Based on DO-Conv, we propose a novel tracking algorithm in Siamese framework (named DOSiam). Extensive experiments conducted on five benchmarks including OTB2015, VOT2016, VOT2018, GOT-10k and VOT2019-RGBT(TIR) show that the proposed DOSiam achieves leading tracking performance with real-time tracking speed at 60 FPS against state-of-the-art trackers.
format Online
Article
Text
id pubmed-9432743
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-94327432022-09-01 Siamese network with a depthwise over-parameterized convolutional layer for visual tracking Wang, Yuanyun Zhang, Wenshuang Zhang, Limin Wang, Jun PLoS One Research Article Visual tracking is a fundamental research task in vision computer. It has broad application prospects, such as military defense and civil security. Visual tracking encounters many challenges in practical application, such as occlusion, fast motion and background clutter. Siamese based trackers achieve superior tracking performance in balanced accuracy and tracking speed. The deep feature extraction with Convolutional Neural Network (CNN) is an essential component in Siamese tracking framework. Although existing trackers take full advantage of deep feature information, the spatial structure and semantic information are not adequately exploited, which are helpful for enhancing target representations. The lack of these spatial and semantic information may lead to tracking drift. In this paper, we design a CNN feature extraction subnetwork based on a Depthwise Over-parameterized Convolutional layer (DO-Conv). A joint convolution method is introduced, namely the conventional and depthwise convolution. The depthwise convolution kernel explores independent channel information, which effectively extracts shallow spatial information and deep semantic information, and discards background information. Based on DO-Conv, we propose a novel tracking algorithm in Siamese framework (named DOSiam). Extensive experiments conducted on five benchmarks including OTB2015, VOT2016, VOT2018, GOT-10k and VOT2019-RGBT(TIR) show that the proposed DOSiam achieves leading tracking performance with real-time tracking speed at 60 FPS against state-of-the-art trackers. Public Library of Science 2022-08-31 /pmc/articles/PMC9432743/ /pubmed/36044439 http://dx.doi.org/10.1371/journal.pone.0273690 Text en © 2022 Wang et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Wang, Yuanyun
Zhang, Wenshuang
Zhang, Limin
Wang, Jun
Siamese network with a depthwise over-parameterized convolutional layer for visual tracking
title Siamese network with a depthwise over-parameterized convolutional layer for visual tracking
title_full Siamese network with a depthwise over-parameterized convolutional layer for visual tracking
title_fullStr Siamese network with a depthwise over-parameterized convolutional layer for visual tracking
title_full_unstemmed Siamese network with a depthwise over-parameterized convolutional layer for visual tracking
title_short Siamese network with a depthwise over-parameterized convolutional layer for visual tracking
title_sort siamese network with a depthwise over-parameterized convolutional layer for visual tracking
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9432743/
https://www.ncbi.nlm.nih.gov/pubmed/36044439
http://dx.doi.org/10.1371/journal.pone.0273690
work_keys_str_mv AT wangyuanyun siamesenetworkwithadepthwiseoverparameterizedconvolutionallayerforvisualtracking
AT zhangwenshuang siamesenetworkwithadepthwiseoverparameterizedconvolutionallayerforvisualtracking
AT zhanglimin siamesenetworkwithadepthwiseoverparameterizedconvolutionallayerforvisualtracking
AT wangjun siamesenetworkwithadepthwiseoverparameterizedconvolutionallayerforvisualtracking