Cargando…
Siamese network with a depthwise over-parameterized convolutional layer for visual tracking
Visual tracking is a fundamental research task in vision computer. It has broad application prospects, such as military defense and civil security. Visual tracking encounters many challenges in practical application, such as occlusion, fast motion and background clutter. Siamese based trackers achie...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9432743/ https://www.ncbi.nlm.nih.gov/pubmed/36044439 http://dx.doi.org/10.1371/journal.pone.0273690 |
_version_ | 1784780455004864512 |
---|---|
author | Wang, Yuanyun Zhang, Wenshuang Zhang, Limin Wang, Jun |
author_facet | Wang, Yuanyun Zhang, Wenshuang Zhang, Limin Wang, Jun |
author_sort | Wang, Yuanyun |
collection | PubMed |
description | Visual tracking is a fundamental research task in vision computer. It has broad application prospects, such as military defense and civil security. Visual tracking encounters many challenges in practical application, such as occlusion, fast motion and background clutter. Siamese based trackers achieve superior tracking performance in balanced accuracy and tracking speed. The deep feature extraction with Convolutional Neural Network (CNN) is an essential component in Siamese tracking framework. Although existing trackers take full advantage of deep feature information, the spatial structure and semantic information are not adequately exploited, which are helpful for enhancing target representations. The lack of these spatial and semantic information may lead to tracking drift. In this paper, we design a CNN feature extraction subnetwork based on a Depthwise Over-parameterized Convolutional layer (DO-Conv). A joint convolution method is introduced, namely the conventional and depthwise convolution. The depthwise convolution kernel explores independent channel information, which effectively extracts shallow spatial information and deep semantic information, and discards background information. Based on DO-Conv, we propose a novel tracking algorithm in Siamese framework (named DOSiam). Extensive experiments conducted on five benchmarks including OTB2015, VOT2016, VOT2018, GOT-10k and VOT2019-RGBT(TIR) show that the proposed DOSiam achieves leading tracking performance with real-time tracking speed at 60 FPS against state-of-the-art trackers. |
format | Online Article Text |
id | pubmed-9432743 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-94327432022-09-01 Siamese network with a depthwise over-parameterized convolutional layer for visual tracking Wang, Yuanyun Zhang, Wenshuang Zhang, Limin Wang, Jun PLoS One Research Article Visual tracking is a fundamental research task in vision computer. It has broad application prospects, such as military defense and civil security. Visual tracking encounters many challenges in practical application, such as occlusion, fast motion and background clutter. Siamese based trackers achieve superior tracking performance in balanced accuracy and tracking speed. The deep feature extraction with Convolutional Neural Network (CNN) is an essential component in Siamese tracking framework. Although existing trackers take full advantage of deep feature information, the spatial structure and semantic information are not adequately exploited, which are helpful for enhancing target representations. The lack of these spatial and semantic information may lead to tracking drift. In this paper, we design a CNN feature extraction subnetwork based on a Depthwise Over-parameterized Convolutional layer (DO-Conv). A joint convolution method is introduced, namely the conventional and depthwise convolution. The depthwise convolution kernel explores independent channel information, which effectively extracts shallow spatial information and deep semantic information, and discards background information. Based on DO-Conv, we propose a novel tracking algorithm in Siamese framework (named DOSiam). Extensive experiments conducted on five benchmarks including OTB2015, VOT2016, VOT2018, GOT-10k and VOT2019-RGBT(TIR) show that the proposed DOSiam achieves leading tracking performance with real-time tracking speed at 60 FPS against state-of-the-art trackers. Public Library of Science 2022-08-31 /pmc/articles/PMC9432743/ /pubmed/36044439 http://dx.doi.org/10.1371/journal.pone.0273690 Text en © 2022 Wang et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Wang, Yuanyun Zhang, Wenshuang Zhang, Limin Wang, Jun Siamese network with a depthwise over-parameterized convolutional layer for visual tracking |
title | Siamese network with a depthwise over-parameterized convolutional layer for visual tracking |
title_full | Siamese network with a depthwise over-parameterized convolutional layer for visual tracking |
title_fullStr | Siamese network with a depthwise over-parameterized convolutional layer for visual tracking |
title_full_unstemmed | Siamese network with a depthwise over-parameterized convolutional layer for visual tracking |
title_short | Siamese network with a depthwise over-parameterized convolutional layer for visual tracking |
title_sort | siamese network with a depthwise over-parameterized convolutional layer for visual tracking |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9432743/ https://www.ncbi.nlm.nih.gov/pubmed/36044439 http://dx.doi.org/10.1371/journal.pone.0273690 |
work_keys_str_mv | AT wangyuanyun siamesenetworkwithadepthwiseoverparameterizedconvolutionallayerforvisualtracking AT zhangwenshuang siamesenetworkwithadepthwiseoverparameterizedconvolutionallayerforvisualtracking AT zhanglimin siamesenetworkwithadepthwiseoverparameterizedconvolutionallayerforvisualtracking AT wangjun siamesenetworkwithadepthwiseoverparameterizedconvolutionallayerforvisualtracking |