Cargando…

Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation

Autonomous systems need to localize and track surrounding objects in 3D space for safe motion planning. As a result, 3D multi-object tracking (MOT) plays a vital role in autonomous navigation. Most MOT methods use a tracking-by-detection pipeline, which includes both the object detection and data as...

Descripción completa

Detalles Bibliográficos
Autores principales: Chen, Can, Zanotti Fragonara, Luca, Tsourdos, Antonios
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8002739/
https://www.ncbi.nlm.nih.gov/pubmed/33803021
http://dx.doi.org/10.3390/s21062113
_version_ 1783671533966721024
author Chen, Can
Zanotti Fragonara, Luca
Tsourdos, Antonios
author_facet Chen, Can
Zanotti Fragonara, Luca
Tsourdos, Antonios
author_sort Chen, Can
collection PubMed
description Autonomous systems need to localize and track surrounding objects in 3D space for safe motion planning. As a result, 3D multi-object tracking (MOT) plays a vital role in autonomous navigation. Most MOT methods use a tracking-by-detection pipeline, which includes both the object detection and data association tasks. However, many approaches detect objects in 2D RGB sequences for tracking, which lacks reliability when localizing objects in 3D space. Furthermore, it is still challenging to learn discriminative features for temporally consistent detection in different frames, and the affinity matrix is typically learned from independent object features without considering the feature interaction between detected objects in the different frames. To settle these problems, we first employ a joint feature extractor to fuse the appearance feature and the motion feature captured from 2D RGB images and 3D point clouds, and then we propose a novel convolutional operation, named RelationConv, to better exploit the correlation between each pair of objects in the adjacent frames and learn a deep affinity matrix for further data association. We finally provide extensive evaluation to reveal that our proposed model achieves state-of-the-art performance on the KITTI tracking benchmark.
format Online
Article
Text
id pubmed-8002739
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-80027392021-03-28 Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation Chen, Can Zanotti Fragonara, Luca Tsourdos, Antonios Sensors (Basel) Article Autonomous systems need to localize and track surrounding objects in 3D space for safe motion planning. As a result, 3D multi-object tracking (MOT) plays a vital role in autonomous navigation. Most MOT methods use a tracking-by-detection pipeline, which includes both the object detection and data association tasks. However, many approaches detect objects in 2D RGB sequences for tracking, which lacks reliability when localizing objects in 3D space. Furthermore, it is still challenging to learn discriminative features for temporally consistent detection in different frames, and the affinity matrix is typically learned from independent object features without considering the feature interaction between detected objects in the different frames. To settle these problems, we first employ a joint feature extractor to fuse the appearance feature and the motion feature captured from 2D RGB images and 3D point clouds, and then we propose a novel convolutional operation, named RelationConv, to better exploit the correlation between each pair of objects in the adjacent frames and learn a deep affinity matrix for further data association. We finally provide extensive evaluation to reveal that our proposed model achieves state-of-the-art performance on the KITTI tracking benchmark. MDPI 2021-03-17 /pmc/articles/PMC8002739/ /pubmed/33803021 http://dx.doi.org/10.3390/s21062113 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Chen, Can
Zanotti Fragonara, Luca
Tsourdos, Antonios
Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation
title Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation
title_full Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation
title_fullStr Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation
title_full_unstemmed Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation
title_short Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation
title_sort relation3dmot: exploiting deep affinity for 3d multi-object tracking from view aggregation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8002739/
https://www.ncbi.nlm.nih.gov/pubmed/33803021
http://dx.doi.org/10.3390/s21062113
work_keys_str_mv AT chencan relation3dmotexploitingdeepaffinityfor3dmultiobjecttrackingfromviewaggregation
AT zanottifragonaraluca relation3dmotexploitingdeepaffinityfor3dmultiobjecttrackingfromviewaggregation
AT tsourdosantonios relation3dmotexploitingdeepaffinityfor3dmultiobjecttrackingfromviewaggregation