Cargando…

Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking

Dictionary learning is a method of acquiring a collection of atoms for subsequent signal representation. Due to its excellent representation ability, dictionary learning has been widely applied in multimedia and computer vision. However, conventional dictionary learning algorithms fail to deal with...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhang, Xiang, Guan, Naiyang, Tao, Dacheng, Qiu, Xiaogang, Luo, Zhigang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2015
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4427315/
https://www.ncbi.nlm.nih.gov/pubmed/25961715
http://dx.doi.org/10.1371/journal.pone.0124685
_version_ 1782370711910744064
author Zhang, Xiang
Guan, Naiyang
Tao, Dacheng
Qiu, Xiaogang
Luo, Zhigang
author_facet Zhang, Xiang
Guan, Naiyang
Tao, Dacheng
Qiu, Xiaogang
Luo, Zhigang
author_sort Zhang, Xiang
collection PubMed
description Dictionary learning is a method of acquiring a collection of atoms for subsequent signal representation. Due to its excellent representation ability, dictionary learning has been widely applied in multimedia and computer vision. However, conventional dictionary learning algorithms fail to deal with multi-modal datasets. In this paper, we propose an online multi-modal robust non-negative dictionary learning (OMRNDL) algorithm to overcome this deficiency. Notably, OMRNDL casts visual tracking as a dictionary learning problem under the particle filter framework and captures the intrinsic knowledge about the target from multiple visual modalities, e.g., pixel intensity and texture information. To this end, OMRNDL adaptively learns an individual dictionary, i.e., template, for each modality from available frames, and then represents new particles over all the learned dictionaries by minimizing the fitting loss of data based on M-estimation. The resultant representation coefficient can be viewed as the common semantic representation of particles across multiple modalities, and can be utilized to track the target. OMRNDL incrementally learns the dictionary and the coefficient of each particle by using multiplicative update rules to respectively guarantee their non-negativity constraints. Experimental results on a popular challenging video benchmark validate the effectiveness of OMRNDL for visual tracking in both quantity and quality.
format Online
Article
Text
id pubmed-4427315
institution National Center for Biotechnology Information
language English
publishDate 2015
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-44273152015-05-21 Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking Zhang, Xiang Guan, Naiyang Tao, Dacheng Qiu, Xiaogang Luo, Zhigang PLoS One Research Article Dictionary learning is a method of acquiring a collection of atoms for subsequent signal representation. Due to its excellent representation ability, dictionary learning has been widely applied in multimedia and computer vision. However, conventional dictionary learning algorithms fail to deal with multi-modal datasets. In this paper, we propose an online multi-modal robust non-negative dictionary learning (OMRNDL) algorithm to overcome this deficiency. Notably, OMRNDL casts visual tracking as a dictionary learning problem under the particle filter framework and captures the intrinsic knowledge about the target from multiple visual modalities, e.g., pixel intensity and texture information. To this end, OMRNDL adaptively learns an individual dictionary, i.e., template, for each modality from available frames, and then represents new particles over all the learned dictionaries by minimizing the fitting loss of data based on M-estimation. The resultant representation coefficient can be viewed as the common semantic representation of particles across multiple modalities, and can be utilized to track the target. OMRNDL incrementally learns the dictionary and the coefficient of each particle by using multiplicative update rules to respectively guarantee their non-negativity constraints. Experimental results on a popular challenging video benchmark validate the effectiveness of OMRNDL for visual tracking in both quantity and quality. Public Library of Science 2015-05-11 /pmc/articles/PMC4427315/ /pubmed/25961715 http://dx.doi.org/10.1371/journal.pone.0124685 Text en © 2015 Zhang et al http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited.
spellingShingle Research Article
Zhang, Xiang
Guan, Naiyang
Tao, Dacheng
Qiu, Xiaogang
Luo, Zhigang
Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking
title Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking
title_full Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking
title_fullStr Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking
title_full_unstemmed Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking
title_short Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking
title_sort online multi-modal robust non-negative dictionary learning for visual tracking
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4427315/
https://www.ncbi.nlm.nih.gov/pubmed/25961715
http://dx.doi.org/10.1371/journal.pone.0124685
work_keys_str_mv AT zhangxiang onlinemultimodalrobustnonnegativedictionarylearningforvisualtracking
AT guannaiyang onlinemultimodalrobustnonnegativedictionarylearningforvisualtracking
AT taodacheng onlinemultimodalrobustnonnegativedictionarylearningforvisualtracking
AT qiuxiaogang onlinemultimodalrobustnonnegativedictionarylearningforvisualtracking
AT luozhigang onlinemultimodalrobustnonnegativedictionarylearningforvisualtracking