Cargando…
Multi-View Fusion-Based 3D Object Detection for Robot Indoor Scene Perception
To autonomously move and operate objects in cluttered indoor environments, a service robot requires the ability of 3D scene perception. Though 3D object detection can provide an object-level environmental description to fill this gap, a robot always encounters incomplete object observation, recurrin...
Autores principales: | , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6806321/ https://www.ncbi.nlm.nih.gov/pubmed/31546674 http://dx.doi.org/10.3390/s19194092 |
_version_ | 1783461603220389888 |
---|---|
author | Wang, Li Li, Ruifeng Sun, Jingwen Liu, Xingxing Zhao, Lijun Seah, Hock Soon Quah, Chee Kwang Tandianus, Budianto |
author_facet | Wang, Li Li, Ruifeng Sun, Jingwen Liu, Xingxing Zhao, Lijun Seah, Hock Soon Quah, Chee Kwang Tandianus, Budianto |
author_sort | Wang, Li |
collection | PubMed |
description | To autonomously move and operate objects in cluttered indoor environments, a service robot requires the ability of 3D scene perception. Though 3D object detection can provide an object-level environmental description to fill this gap, a robot always encounters incomplete object observation, recurring detections of the same object, error in detection, or intersection between objects when conducting detection continuously in a cluttered room. To solve these problems, we propose a two-stage 3D object detection algorithm which is to fuse multiple views of 3D object point clouds in the first stage and to eliminate unreasonable and intersection detections in the second stage. For each view, the robot performs a 2D object semantic segmentation and obtains 3D object point clouds. Then, an unsupervised segmentation method called Locally Convex Connected Patches (LCCP) is utilized to segment the object accurately from the background. Subsequently, the Manhattan Frame estimation is implemented to calculate the main orientation of the object and subsequently, the 3D object bounding box can be obtained. To deal with the detected objects in multiple views, we construct an object database and propose an object fusion criterion to maintain it automatically. Thus, the same object observed in multi-view is fused together and a more accurate bounding box can be calculated. Finally, we propose an object filtering approach based on prior knowledge to remove incorrect and intersecting objects in the object dataset. Experiments are carried out on both SceneNN dataset and a real indoor environment to verify the stability and accuracy of 3D semantic segmentation and bounding box detection of the object with multi-view fusion. |
format | Online Article Text |
id | pubmed-6806321 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-68063212019-11-07 Multi-View Fusion-Based 3D Object Detection for Robot Indoor Scene Perception Wang, Li Li, Ruifeng Sun, Jingwen Liu, Xingxing Zhao, Lijun Seah, Hock Soon Quah, Chee Kwang Tandianus, Budianto Sensors (Basel) Article To autonomously move and operate objects in cluttered indoor environments, a service robot requires the ability of 3D scene perception. Though 3D object detection can provide an object-level environmental description to fill this gap, a robot always encounters incomplete object observation, recurring detections of the same object, error in detection, or intersection between objects when conducting detection continuously in a cluttered room. To solve these problems, we propose a two-stage 3D object detection algorithm which is to fuse multiple views of 3D object point clouds in the first stage and to eliminate unreasonable and intersection detections in the second stage. For each view, the robot performs a 2D object semantic segmentation and obtains 3D object point clouds. Then, an unsupervised segmentation method called Locally Convex Connected Patches (LCCP) is utilized to segment the object accurately from the background. Subsequently, the Manhattan Frame estimation is implemented to calculate the main orientation of the object and subsequently, the 3D object bounding box can be obtained. To deal with the detected objects in multiple views, we construct an object database and propose an object fusion criterion to maintain it automatically. Thus, the same object observed in multi-view is fused together and a more accurate bounding box can be calculated. Finally, we propose an object filtering approach based on prior knowledge to remove incorrect and intersecting objects in the object dataset. Experiments are carried out on both SceneNN dataset and a real indoor environment to verify the stability and accuracy of 3D semantic segmentation and bounding box detection of the object with multi-view fusion. MDPI 2019-09-21 /pmc/articles/PMC6806321/ /pubmed/31546674 http://dx.doi.org/10.3390/s19194092 Text en © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Wang, Li Li, Ruifeng Sun, Jingwen Liu, Xingxing Zhao, Lijun Seah, Hock Soon Quah, Chee Kwang Tandianus, Budianto Multi-View Fusion-Based 3D Object Detection for Robot Indoor Scene Perception |
title | Multi-View Fusion-Based 3D Object Detection for Robot Indoor Scene Perception |
title_full | Multi-View Fusion-Based 3D Object Detection for Robot Indoor Scene Perception |
title_fullStr | Multi-View Fusion-Based 3D Object Detection for Robot Indoor Scene Perception |
title_full_unstemmed | Multi-View Fusion-Based 3D Object Detection for Robot Indoor Scene Perception |
title_short | Multi-View Fusion-Based 3D Object Detection for Robot Indoor Scene Perception |
title_sort | multi-view fusion-based 3d object detection for robot indoor scene perception |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6806321/ https://www.ncbi.nlm.nih.gov/pubmed/31546674 http://dx.doi.org/10.3390/s19194092 |
work_keys_str_mv | AT wangli multiviewfusionbased3dobjectdetectionforrobotindoorsceneperception AT liruifeng multiviewfusionbased3dobjectdetectionforrobotindoorsceneperception AT sunjingwen multiviewfusionbased3dobjectdetectionforrobotindoorsceneperception AT liuxingxing multiviewfusionbased3dobjectdetectionforrobotindoorsceneperception AT zhaolijun multiviewfusionbased3dobjectdetectionforrobotindoorsceneperception AT seahhocksoon multiviewfusionbased3dobjectdetectionforrobotindoorsceneperception AT quahcheekwang multiviewfusionbased3dobjectdetectionforrobotindoorsceneperception AT tandianusbudianto multiviewfusionbased3dobjectdetectionforrobotindoorsceneperception |