Cargando…
Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation
Autonomous systems need to localize and track surrounding objects in 3D space for safe motion planning. As a result, 3D multi-object tracking (MOT) plays a vital role in autonomous navigation. Most MOT methods use a tracking-by-detection pipeline, which includes both the object detection and data as...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8002739/ https://www.ncbi.nlm.nih.gov/pubmed/33803021 http://dx.doi.org/10.3390/s21062113 |
_version_ | 1783671533966721024 |
---|---|
author | Chen, Can Zanotti Fragonara, Luca Tsourdos, Antonios |
author_facet | Chen, Can Zanotti Fragonara, Luca Tsourdos, Antonios |
author_sort | Chen, Can |
collection | PubMed |
description | Autonomous systems need to localize and track surrounding objects in 3D space for safe motion planning. As a result, 3D multi-object tracking (MOT) plays a vital role in autonomous navigation. Most MOT methods use a tracking-by-detection pipeline, which includes both the object detection and data association tasks. However, many approaches detect objects in 2D RGB sequences for tracking, which lacks reliability when localizing objects in 3D space. Furthermore, it is still challenging to learn discriminative features for temporally consistent detection in different frames, and the affinity matrix is typically learned from independent object features without considering the feature interaction between detected objects in the different frames. To settle these problems, we first employ a joint feature extractor to fuse the appearance feature and the motion feature captured from 2D RGB images and 3D point clouds, and then we propose a novel convolutional operation, named RelationConv, to better exploit the correlation between each pair of objects in the adjacent frames and learn a deep affinity matrix for further data association. We finally provide extensive evaluation to reveal that our proposed model achieves state-of-the-art performance on the KITTI tracking benchmark. |
format | Online Article Text |
id | pubmed-8002739 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-80027392021-03-28 Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation Chen, Can Zanotti Fragonara, Luca Tsourdos, Antonios Sensors (Basel) Article Autonomous systems need to localize and track surrounding objects in 3D space for safe motion planning. As a result, 3D multi-object tracking (MOT) plays a vital role in autonomous navigation. Most MOT methods use a tracking-by-detection pipeline, which includes both the object detection and data association tasks. However, many approaches detect objects in 2D RGB sequences for tracking, which lacks reliability when localizing objects in 3D space. Furthermore, it is still challenging to learn discriminative features for temporally consistent detection in different frames, and the affinity matrix is typically learned from independent object features without considering the feature interaction between detected objects in the different frames. To settle these problems, we first employ a joint feature extractor to fuse the appearance feature and the motion feature captured from 2D RGB images and 3D point clouds, and then we propose a novel convolutional operation, named RelationConv, to better exploit the correlation between each pair of objects in the adjacent frames and learn a deep affinity matrix for further data association. We finally provide extensive evaluation to reveal that our proposed model achieves state-of-the-art performance on the KITTI tracking benchmark. MDPI 2021-03-17 /pmc/articles/PMC8002739/ /pubmed/33803021 http://dx.doi.org/10.3390/s21062113 Text en © 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Chen, Can Zanotti Fragonara, Luca Tsourdos, Antonios Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation |
title | Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation |
title_full | Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation |
title_fullStr | Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation |
title_full_unstemmed | Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation |
title_short | Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking from View Aggregation |
title_sort | relation3dmot: exploiting deep affinity for 3d multi-object tracking from view aggregation |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8002739/ https://www.ncbi.nlm.nih.gov/pubmed/33803021 http://dx.doi.org/10.3390/s21062113 |
work_keys_str_mv | AT chencan relation3dmotexploitingdeepaffinityfor3dmultiobjecttrackingfromviewaggregation AT zanottifragonaraluca relation3dmotexploitingdeepaffinityfor3dmultiobjecttrackingfromviewaggregation AT tsourdosantonios relation3dmotexploitingdeepaffinityfor3dmultiobjecttrackingfromviewaggregation |