Cargando…

Fast and Robust Visual Tracking with Few-Iteration Meta-Learning

Visual object tracking has been a major research topic in the field of computer vision for many years. Object tracking aims to identify and localize objects of interest in subsequent frames, given the bounding box of the first frame. In addition, the object-tracking algorithms are also required to h...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Zhenxin, Zhang, Xuande, Xu, Long, Zhang, Weiqiang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9370940/
https://www.ncbi.nlm.nih.gov/pubmed/35957383
http://dx.doi.org/10.3390/s22155826
_version_ 1784766972147269632
author Li, Zhenxin
Zhang, Xuande
Xu, Long
Zhang, Weiqiang
author_facet Li, Zhenxin
Zhang, Xuande
Xu, Long
Zhang, Weiqiang
author_sort Li, Zhenxin
collection PubMed
description Visual object tracking has been a major research topic in the field of computer vision for many years. Object tracking aims to identify and localize objects of interest in subsequent frames, given the bounding box of the first frame. In addition, the object-tracking algorithms are also required to have robustness and real-time performance. These requirements create some unique challenges, which can easily become overfitting if given a very small training dataset of objects during offline training. On the other hand, if there are too many iterations in the model-optimization process during offline training or in the model-update process during online tracking, it will cause the problem of poor real-time performance. We address these problems by introducing a meta-learning method based on fast optimization. Our proposed tracking architecture mainly contains two parts, one is the base learner and the other is the meta learner. The base learner is primarily a target and background classifier, in addition, there is an object bounding box prediction regression network. The primary goal of a meta learner based on the transformer is to learn the representations used by the classifier. The accuracy of our proposed algorithm on OTB2015 and LaSOT is 0.930 and 0.688, respectively. Moreover, it performs well on VOT2018 and GOT-10k datasets. Combined with the comparative experiments on real-time performance, our algorithm is fast and robust.
format Online
Article
Text
id pubmed-9370940
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-93709402022-08-12 Fast and Robust Visual Tracking with Few-Iteration Meta-Learning Li, Zhenxin Zhang, Xuande Xu, Long Zhang, Weiqiang Sensors (Basel) Article Visual object tracking has been a major research topic in the field of computer vision for many years. Object tracking aims to identify and localize objects of interest in subsequent frames, given the bounding box of the first frame. In addition, the object-tracking algorithms are also required to have robustness and real-time performance. These requirements create some unique challenges, which can easily become overfitting if given a very small training dataset of objects during offline training. On the other hand, if there are too many iterations in the model-optimization process during offline training or in the model-update process during online tracking, it will cause the problem of poor real-time performance. We address these problems by introducing a meta-learning method based on fast optimization. Our proposed tracking architecture mainly contains two parts, one is the base learner and the other is the meta learner. The base learner is primarily a target and background classifier, in addition, there is an object bounding box prediction regression network. The primary goal of a meta learner based on the transformer is to learn the representations used by the classifier. The accuracy of our proposed algorithm on OTB2015 and LaSOT is 0.930 and 0.688, respectively. Moreover, it performs well on VOT2018 and GOT-10k datasets. Combined with the comparative experiments on real-time performance, our algorithm is fast and robust. MDPI 2022-08-04 /pmc/articles/PMC9370940/ /pubmed/35957383 http://dx.doi.org/10.3390/s22155826 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Li, Zhenxin
Zhang, Xuande
Xu, Long
Zhang, Weiqiang
Fast and Robust Visual Tracking with Few-Iteration Meta-Learning
title Fast and Robust Visual Tracking with Few-Iteration Meta-Learning
title_full Fast and Robust Visual Tracking with Few-Iteration Meta-Learning
title_fullStr Fast and Robust Visual Tracking with Few-Iteration Meta-Learning
title_full_unstemmed Fast and Robust Visual Tracking with Few-Iteration Meta-Learning
title_short Fast and Robust Visual Tracking with Few-Iteration Meta-Learning
title_sort fast and robust visual tracking with few-iteration meta-learning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9370940/
https://www.ncbi.nlm.nih.gov/pubmed/35957383
http://dx.doi.org/10.3390/s22155826
work_keys_str_mv AT lizhenxin fastandrobustvisualtrackingwithfewiterationmetalearning
AT zhangxuande fastandrobustvisualtrackingwithfewiterationmetalearning
AT xulong fastandrobustvisualtrackingwithfewiterationmetalearning
AT zhangweiqiang fastandrobustvisualtrackingwithfewiterationmetalearning