Cargando…
Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking
Dictionary learning is a method of acquiring a collection of atoms for subsequent signal representation. Due to its excellent representation ability, dictionary learning has been widely applied in multimedia and computer vision. However, conventional dictionary learning algorithms fail to deal with...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2015
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4427315/ https://www.ncbi.nlm.nih.gov/pubmed/25961715 http://dx.doi.org/10.1371/journal.pone.0124685 |
_version_ | 1782370711910744064 |
---|---|
author | Zhang, Xiang Guan, Naiyang Tao, Dacheng Qiu, Xiaogang Luo, Zhigang |
author_facet | Zhang, Xiang Guan, Naiyang Tao, Dacheng Qiu, Xiaogang Luo, Zhigang |
author_sort | Zhang, Xiang |
collection | PubMed |
description | Dictionary learning is a method of acquiring a collection of atoms for subsequent signal representation. Due to its excellent representation ability, dictionary learning has been widely applied in multimedia and computer vision. However, conventional dictionary learning algorithms fail to deal with multi-modal datasets. In this paper, we propose an online multi-modal robust non-negative dictionary learning (OMRNDL) algorithm to overcome this deficiency. Notably, OMRNDL casts visual tracking as a dictionary learning problem under the particle filter framework and captures the intrinsic knowledge about the target from multiple visual modalities, e.g., pixel intensity and texture information. To this end, OMRNDL adaptively learns an individual dictionary, i.e., template, for each modality from available frames, and then represents new particles over all the learned dictionaries by minimizing the fitting loss of data based on M-estimation. The resultant representation coefficient can be viewed as the common semantic representation of particles across multiple modalities, and can be utilized to track the target. OMRNDL incrementally learns the dictionary and the coefficient of each particle by using multiplicative update rules to respectively guarantee their non-negativity constraints. Experimental results on a popular challenging video benchmark validate the effectiveness of OMRNDL for visual tracking in both quantity and quality. |
format | Online Article Text |
id | pubmed-4427315 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2015 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-44273152015-05-21 Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking Zhang, Xiang Guan, Naiyang Tao, Dacheng Qiu, Xiaogang Luo, Zhigang PLoS One Research Article Dictionary learning is a method of acquiring a collection of atoms for subsequent signal representation. Due to its excellent representation ability, dictionary learning has been widely applied in multimedia and computer vision. However, conventional dictionary learning algorithms fail to deal with multi-modal datasets. In this paper, we propose an online multi-modal robust non-negative dictionary learning (OMRNDL) algorithm to overcome this deficiency. Notably, OMRNDL casts visual tracking as a dictionary learning problem under the particle filter framework and captures the intrinsic knowledge about the target from multiple visual modalities, e.g., pixel intensity and texture information. To this end, OMRNDL adaptively learns an individual dictionary, i.e., template, for each modality from available frames, and then represents new particles over all the learned dictionaries by minimizing the fitting loss of data based on M-estimation. The resultant representation coefficient can be viewed as the common semantic representation of particles across multiple modalities, and can be utilized to track the target. OMRNDL incrementally learns the dictionary and the coefficient of each particle by using multiplicative update rules to respectively guarantee their non-negativity constraints. Experimental results on a popular challenging video benchmark validate the effectiveness of OMRNDL for visual tracking in both quantity and quality. Public Library of Science 2015-05-11 /pmc/articles/PMC4427315/ /pubmed/25961715 http://dx.doi.org/10.1371/journal.pone.0124685 Text en © 2015 Zhang et al http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited. |
spellingShingle | Research Article Zhang, Xiang Guan, Naiyang Tao, Dacheng Qiu, Xiaogang Luo, Zhigang Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking |
title | Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking |
title_full | Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking |
title_fullStr | Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking |
title_full_unstemmed | Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking |
title_short | Online Multi-Modal Robust Non-Negative Dictionary Learning for Visual Tracking |
title_sort | online multi-modal robust non-negative dictionary learning for visual tracking |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4427315/ https://www.ncbi.nlm.nih.gov/pubmed/25961715 http://dx.doi.org/10.1371/journal.pone.0124685 |
work_keys_str_mv | AT zhangxiang onlinemultimodalrobustnonnegativedictionarylearningforvisualtracking AT guannaiyang onlinemultimodalrobustnonnegativedictionarylearningforvisualtracking AT taodacheng onlinemultimodalrobustnonnegativedictionarylearningforvisualtracking AT qiuxiaogang onlinemultimodalrobustnonnegativedictionarylearningforvisualtracking AT luozhigang onlinemultimodalrobustnonnegativedictionarylearningforvisualtracking |