Cargando…
Point Pair Feature-Based Pose Estimation with Multiple Edge Appearance Models (PPF-MEAM) for Robotic Bin Picking
Automation of the bin picking task with robots entails the key step of pose estimation, which identifies and locates objects so that the robot can pick and manipulate the object in an accurate and reliable way. This paper proposes a novel point pair feature-based descriptor named Boundary-to-Boundar...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6111311/ https://www.ncbi.nlm.nih.gov/pubmed/30126220 http://dx.doi.org/10.3390/s18082719 |
_version_ | 1783350631447134208 |
---|---|
author | Liu, Diyi Arai, Shogo Miao, Jiaqi Kinugawa, Jun Wang, Zhao Kosuge, Kazuhiro |
author_facet | Liu, Diyi Arai, Shogo Miao, Jiaqi Kinugawa, Jun Wang, Zhao Kosuge, Kazuhiro |
author_sort | Liu, Diyi |
collection | PubMed |
description | Automation of the bin picking task with robots entails the key step of pose estimation, which identifies and locates objects so that the robot can pick and manipulate the object in an accurate and reliable way. This paper proposes a novel point pair feature-based descriptor named Boundary-to-Boundary-using-Tangent-Line (B2B-TL) to estimate the pose of industrial parts including some parts whose point clouds lack key details, for example, the point cloud of the ridges of a part. The proposed descriptor utilizes the 3D point cloud data and 2D image data of the scene simultaneously, and the 2D image data could compensate the missing key details of the point cloud. Based on the descriptor B2B-TL, Multiple Edge Appearance Models (MEAM), a method using multiple models to describe the target object, is proposed to increase the recognition rate and reduce the computation time. A novel pipeline of an online computation process is presented to take advantage of B2B-TL and MEAM. Our algorithm is evaluated against synthetic and real scenes and implemented in a bin picking system. The experimental results show that our method is sufficiently accurate for a robot to grasp industrial parts and is fast enough to be used in a real factory environment. |
format | Online Article Text |
id | pubmed-6111311 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-61113112018-08-30 Point Pair Feature-Based Pose Estimation with Multiple Edge Appearance Models (PPF-MEAM) for Robotic Bin Picking Liu, Diyi Arai, Shogo Miao, Jiaqi Kinugawa, Jun Wang, Zhao Kosuge, Kazuhiro Sensors (Basel) Article Automation of the bin picking task with robots entails the key step of pose estimation, which identifies and locates objects so that the robot can pick and manipulate the object in an accurate and reliable way. This paper proposes a novel point pair feature-based descriptor named Boundary-to-Boundary-using-Tangent-Line (B2B-TL) to estimate the pose of industrial parts including some parts whose point clouds lack key details, for example, the point cloud of the ridges of a part. The proposed descriptor utilizes the 3D point cloud data and 2D image data of the scene simultaneously, and the 2D image data could compensate the missing key details of the point cloud. Based on the descriptor B2B-TL, Multiple Edge Appearance Models (MEAM), a method using multiple models to describe the target object, is proposed to increase the recognition rate and reduce the computation time. A novel pipeline of an online computation process is presented to take advantage of B2B-TL and MEAM. Our algorithm is evaluated against synthetic and real scenes and implemented in a bin picking system. The experimental results show that our method is sufficiently accurate for a robot to grasp industrial parts and is fast enough to be used in a real factory environment. MDPI 2018-08-18 /pmc/articles/PMC6111311/ /pubmed/30126220 http://dx.doi.org/10.3390/s18082719 Text en © 2018 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Liu, Diyi Arai, Shogo Miao, Jiaqi Kinugawa, Jun Wang, Zhao Kosuge, Kazuhiro Point Pair Feature-Based Pose Estimation with Multiple Edge Appearance Models (PPF-MEAM) for Robotic Bin Picking |
title | Point Pair Feature-Based Pose Estimation with Multiple Edge Appearance Models (PPF-MEAM) for Robotic Bin Picking |
title_full | Point Pair Feature-Based Pose Estimation with Multiple Edge Appearance Models (PPF-MEAM) for Robotic Bin Picking |
title_fullStr | Point Pair Feature-Based Pose Estimation with Multiple Edge Appearance Models (PPF-MEAM) for Robotic Bin Picking |
title_full_unstemmed | Point Pair Feature-Based Pose Estimation with Multiple Edge Appearance Models (PPF-MEAM) for Robotic Bin Picking |
title_short | Point Pair Feature-Based Pose Estimation with Multiple Edge Appearance Models (PPF-MEAM) for Robotic Bin Picking |
title_sort | point pair feature-based pose estimation with multiple edge appearance models (ppf-meam) for robotic bin picking |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6111311/ https://www.ncbi.nlm.nih.gov/pubmed/30126220 http://dx.doi.org/10.3390/s18082719 |
work_keys_str_mv | AT liudiyi pointpairfeaturebasedposeestimationwithmultipleedgeappearancemodelsppfmeamforroboticbinpicking AT araishogo pointpairfeaturebasedposeestimationwithmultipleedgeappearancemodelsppfmeamforroboticbinpicking AT miaojiaqi pointpairfeaturebasedposeestimationwithmultipleedgeappearancemodelsppfmeamforroboticbinpicking AT kinugawajun pointpairfeaturebasedposeestimationwithmultipleedgeappearancemodelsppfmeamforroboticbinpicking AT wangzhao pointpairfeaturebasedposeestimationwithmultipleedgeappearancemodelsppfmeamforroboticbinpicking AT kosugekazuhiro pointpairfeaturebasedposeestimationwithmultipleedgeappearancemodelsppfmeamforroboticbinpicking |