Cargando…

Anti-Noise 3D Object Detection of Multimodal Feature Attention Fusion Based on PV-RCNN

3D object detection methods based on camera and LiDAR fusion are susceptible to environmental noise. Due to the mismatch of physical characteristics of the two sensors, the feature vectors encoded by the feature layer are in different feature spaces. This leads to the problem of feature information...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhu, Yuan, Xu, Ruidong, An, Hao, Tao, Chongben, Lu, Ke
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9823336/
https://www.ncbi.nlm.nih.gov/pubmed/36616829
http://dx.doi.org/10.3390/s23010233
_version_ 1784866134546186240
author Zhu, Yuan
Xu, Ruidong
An, Hao
Tao, Chongben
Lu, Ke
author_facet Zhu, Yuan
Xu, Ruidong
An, Hao
Tao, Chongben
Lu, Ke
author_sort Zhu, Yuan
collection PubMed
description 3D object detection methods based on camera and LiDAR fusion are susceptible to environmental noise. Due to the mismatch of physical characteristics of the two sensors, the feature vectors encoded by the feature layer are in different feature spaces. This leads to the problem of feature information deviation, which has an impact on detection performance. To address this problem, a point-guided feature abstract method is presented to fuse the camera and LiDAR at first. The extracted image features and point cloud features are aggregated to keypoints for enhancing information redundancy. Second, the proposed multimodal feature attention (MFA) mechanism is used to achieve adaptive fusion of point cloud features and image features with information from multiple feature spaces. Finally, a projection-based farthest point sampling (P-FPS) is proposed to downsample the raw point cloud, which can project more keypoints onto the close object and improve the sampling rate of the point-guided image features. The 3D bounding boxes of the object is obtained by the region of interest (ROI) pooling layer and the fully connected layer. The proposed 3D object detection algorithm is evaluated on three different datasets, and the proposed algorithm achieved better detection performance and robustness when the image and point cloud data contain rain noise. The test results on a physical test platform further validate the effectiveness of the algorithm.
format Online
Article
Text
id pubmed-9823336
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-98233362023-01-08 Anti-Noise 3D Object Detection of Multimodal Feature Attention Fusion Based on PV-RCNN Zhu, Yuan Xu, Ruidong An, Hao Tao, Chongben Lu, Ke Sensors (Basel) Article 3D object detection methods based on camera and LiDAR fusion are susceptible to environmental noise. Due to the mismatch of physical characteristics of the two sensors, the feature vectors encoded by the feature layer are in different feature spaces. This leads to the problem of feature information deviation, which has an impact on detection performance. To address this problem, a point-guided feature abstract method is presented to fuse the camera and LiDAR at first. The extracted image features and point cloud features are aggregated to keypoints for enhancing information redundancy. Second, the proposed multimodal feature attention (MFA) mechanism is used to achieve adaptive fusion of point cloud features and image features with information from multiple feature spaces. Finally, a projection-based farthest point sampling (P-FPS) is proposed to downsample the raw point cloud, which can project more keypoints onto the close object and improve the sampling rate of the point-guided image features. The 3D bounding boxes of the object is obtained by the region of interest (ROI) pooling layer and the fully connected layer. The proposed 3D object detection algorithm is evaluated on three different datasets, and the proposed algorithm achieved better detection performance and robustness when the image and point cloud data contain rain noise. The test results on a physical test platform further validate the effectiveness of the algorithm. MDPI 2022-12-26 /pmc/articles/PMC9823336/ /pubmed/36616829 http://dx.doi.org/10.3390/s23010233 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhu, Yuan
Xu, Ruidong
An, Hao
Tao, Chongben
Lu, Ke
Anti-Noise 3D Object Detection of Multimodal Feature Attention Fusion Based on PV-RCNN
title Anti-Noise 3D Object Detection of Multimodal Feature Attention Fusion Based on PV-RCNN
title_full Anti-Noise 3D Object Detection of Multimodal Feature Attention Fusion Based on PV-RCNN
title_fullStr Anti-Noise 3D Object Detection of Multimodal Feature Attention Fusion Based on PV-RCNN
title_full_unstemmed Anti-Noise 3D Object Detection of Multimodal Feature Attention Fusion Based on PV-RCNN
title_short Anti-Noise 3D Object Detection of Multimodal Feature Attention Fusion Based on PV-RCNN
title_sort anti-noise 3d object detection of multimodal feature attention fusion based on pv-rcnn
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9823336/
https://www.ncbi.nlm.nih.gov/pubmed/36616829
http://dx.doi.org/10.3390/s23010233
work_keys_str_mv AT zhuyuan antinoise3dobjectdetectionofmultimodalfeatureattentionfusionbasedonpvrcnn
AT xuruidong antinoise3dobjectdetectionofmultimodalfeatureattentionfusionbasedonpvrcnn
AT anhao antinoise3dobjectdetectionofmultimodalfeatureattentionfusionbasedonpvrcnn
AT taochongben antinoise3dobjectdetectionofmultimodalfeatureattentionfusionbasedonpvrcnn
AT luke antinoise3dobjectdetectionofmultimodalfeatureattentionfusionbasedonpvrcnn