Cargando…

Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature

Video-based person re-identification is an important task with the challenges of lighting variation, low-resolution images, background clutter, occlusion, and human appearance similarity in the multi-camera visual sensor networks. In this paper, we propose a video-based person re-identification meth...

Descripción completa

Detalles Bibliográficos
Autores principales: Sun, Rui, Huang, Qiheng, Xia, Miaomiao, Zhang, Jun
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6263398/
https://www.ncbi.nlm.nih.gov/pubmed/30380623
http://dx.doi.org/10.3390/s18113669
_version_ 1783375285479014400
author Sun, Rui
Huang, Qiheng
Xia, Miaomiao
Zhang, Jun
author_facet Sun, Rui
Huang, Qiheng
Xia, Miaomiao
Zhang, Jun
author_sort Sun, Rui
collection PubMed
description Video-based person re-identification is an important task with the challenges of lighting variation, low-resolution images, background clutter, occlusion, and human appearance similarity in the multi-camera visual sensor networks. In this paper, we propose a video-based person re-identification method called the end-to-end learning architecture with hybrid deep appearance-temporal feature. It can learn the appearance features of pivotal frames, the temporal features, and the independent distance metric of different features. This architecture consists of two-stream deep feature structure and two Siamese networks. For the first-stream structure, we propose the Two-branch Appearance Feature (TAF) sub-structure to obtain the appearance information of persons, and used one of the two Siamese networks to learn the similarity of appearance features of a pairwise person. To utilize the temporal information, we designed the second-stream structure that consisting of the Optical flow Temporal Feature (OTF) sub-structure and another Siamese network, to learn the person’s temporal features and the distances of pairwise features. In addition, we select the pivotal frames of video as inputs to the Inception-V3 network on the Two-branch Appearance Feature sub-structure, and employ the salience-learning fusion layer to fuse the learned global and local appearance features. Extensive experimental results on the PRID2011, iLIDS-VID, and Motion Analysis and Re-identification Set (MARS) datasets showed that the respective proposed architectures reached 79%, 59% and 72% at Rank-1 and had advantages over state-of-the-art algorithms. Meanwhile, it also improved the feature representation ability of persons.
format Online
Article
Text
id pubmed-6263398
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-62633982018-12-12 Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature Sun, Rui Huang, Qiheng Xia, Miaomiao Zhang, Jun Sensors (Basel) Article Video-based person re-identification is an important task with the challenges of lighting variation, low-resolution images, background clutter, occlusion, and human appearance similarity in the multi-camera visual sensor networks. In this paper, we propose a video-based person re-identification method called the end-to-end learning architecture with hybrid deep appearance-temporal feature. It can learn the appearance features of pivotal frames, the temporal features, and the independent distance metric of different features. This architecture consists of two-stream deep feature structure and two Siamese networks. For the first-stream structure, we propose the Two-branch Appearance Feature (TAF) sub-structure to obtain the appearance information of persons, and used one of the two Siamese networks to learn the similarity of appearance features of a pairwise person. To utilize the temporal information, we designed the second-stream structure that consisting of the Optical flow Temporal Feature (OTF) sub-structure and another Siamese network, to learn the person’s temporal features and the distances of pairwise features. In addition, we select the pivotal frames of video as inputs to the Inception-V3 network on the Two-branch Appearance Feature sub-structure, and employ the salience-learning fusion layer to fuse the learned global and local appearance features. Extensive experimental results on the PRID2011, iLIDS-VID, and Motion Analysis and Re-identification Set (MARS) datasets showed that the respective proposed architectures reached 79%, 59% and 72% at Rank-1 and had advantages over state-of-the-art algorithms. Meanwhile, it also improved the feature representation ability of persons. MDPI 2018-10-29 /pmc/articles/PMC6263398/ /pubmed/30380623 http://dx.doi.org/10.3390/s18113669 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Sun, Rui
Huang, Qiheng
Xia, Miaomiao
Zhang, Jun
Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature
title Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature
title_full Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature
title_fullStr Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature
title_full_unstemmed Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature
title_short Video-Based Person Re-Identification by an End-To-End Learning Architecture with Hybrid Deep Appearance-Temporal Feature
title_sort video-based person re-identification by an end-to-end learning architecture with hybrid deep appearance-temporal feature
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6263398/
https://www.ncbi.nlm.nih.gov/pubmed/30380623
http://dx.doi.org/10.3390/s18113669
work_keys_str_mv AT sunrui videobasedpersonreidentificationbyanendtoendlearningarchitecturewithhybriddeepappearancetemporalfeature
AT huangqiheng videobasedpersonreidentificationbyanendtoendlearningarchitecturewithhybriddeepappearancetemporalfeature
AT xiamiaomiao videobasedpersonreidentificationbyanendtoendlearningarchitecturewithhybriddeepappearancetemporalfeature
AT zhangjun videobasedpersonreidentificationbyanendtoendlearningarchitecturewithhybriddeepappearancetemporalfeature