Cargando…
Object Pose Estimation Using Edge Images Synthesized from Shape Information
This paper presents a method for estimating the six Degrees of Freedom (6DoF) pose of texture-less objects from a monocular image by using edge information. The deep learning-based pose estimation method needs a large dataset containing pairs of an image and ground truth pose of objects. To alleviat...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9781246/ https://www.ncbi.nlm.nih.gov/pubmed/36559978 http://dx.doi.org/10.3390/s22249610 |
_version_ | 1784857026926477312 |
---|---|
author | Moteki, Atsunori Saito, Hideo |
author_facet | Moteki, Atsunori Saito, Hideo |
author_sort | Moteki, Atsunori |
collection | PubMed |
description | This paper presents a method for estimating the six Degrees of Freedom (6DoF) pose of texture-less objects from a monocular image by using edge information. The deep learning-based pose estimation method needs a large dataset containing pairs of an image and ground truth pose of objects. To alleviate the cost of collecting a dataset, we focus on the method using a dataset made by computer graphics (CG). This simulation-based method prepares a thousand images by rendering the computer-aided design (CAD) data of the object and trains a deep-learning model. As an inference stage, a monocular RGB image is entered into the model, and the object’s pose is estimated. The representative simulation-based method, Pose Interpreter Networks, uses silhouette images as the input, thereby enabling common feature (contour) extraction from RGB and CG images. However, estimating rotation parameters is less accurate. To overcome this problem, we propose a method to use edge information extracted from the object’s ridgelines for training the deep learning model. Since edge distribution changes largely according to the pose, the estimation of rotation parameters becomes more robust. Through an experiment with simulation data, we quantitatively proved the accuracy improvement compared to the previous method (error rate decreases at a certain condition are translation 22.9% and rotation: 43.4%). Moreover, through an experiment with physical data, we clarified the issues of this method and proposed an effective solution by fine-tuning (error rate decrease at a certain condition are translation 20.1% and rotation 57.7%). |
format | Online Article Text |
id | pubmed-9781246 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-97812462022-12-24 Object Pose Estimation Using Edge Images Synthesized from Shape Information Moteki, Atsunori Saito, Hideo Sensors (Basel) Article This paper presents a method for estimating the six Degrees of Freedom (6DoF) pose of texture-less objects from a monocular image by using edge information. The deep learning-based pose estimation method needs a large dataset containing pairs of an image and ground truth pose of objects. To alleviate the cost of collecting a dataset, we focus on the method using a dataset made by computer graphics (CG). This simulation-based method prepares a thousand images by rendering the computer-aided design (CAD) data of the object and trains a deep-learning model. As an inference stage, a monocular RGB image is entered into the model, and the object’s pose is estimated. The representative simulation-based method, Pose Interpreter Networks, uses silhouette images as the input, thereby enabling common feature (contour) extraction from RGB and CG images. However, estimating rotation parameters is less accurate. To overcome this problem, we propose a method to use edge information extracted from the object’s ridgelines for training the deep learning model. Since edge distribution changes largely according to the pose, the estimation of rotation parameters becomes more robust. Through an experiment with simulation data, we quantitatively proved the accuracy improvement compared to the previous method (error rate decreases at a certain condition are translation 22.9% and rotation: 43.4%). Moreover, through an experiment with physical data, we clarified the issues of this method and proposed an effective solution by fine-tuning (error rate decrease at a certain condition are translation 20.1% and rotation 57.7%). MDPI 2022-12-08 /pmc/articles/PMC9781246/ /pubmed/36559978 http://dx.doi.org/10.3390/s22249610 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Moteki, Atsunori Saito, Hideo Object Pose Estimation Using Edge Images Synthesized from Shape Information |
title | Object Pose Estimation Using Edge Images Synthesized from Shape Information |
title_full | Object Pose Estimation Using Edge Images Synthesized from Shape Information |
title_fullStr | Object Pose Estimation Using Edge Images Synthesized from Shape Information |
title_full_unstemmed | Object Pose Estimation Using Edge Images Synthesized from Shape Information |
title_short | Object Pose Estimation Using Edge Images Synthesized from Shape Information |
title_sort | object pose estimation using edge images synthesized from shape information |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9781246/ https://www.ncbi.nlm.nih.gov/pubmed/36559978 http://dx.doi.org/10.3390/s22249610 |
work_keys_str_mv | AT motekiatsunori objectposeestimationusingedgeimagessynthesizedfromshapeinformation AT saitohideo objectposeestimationusingedgeimagessynthesizedfromshapeinformation |