Cargando…

Object Pose Estimation Using Edge Images Synthesized from Shape Information

This paper presents a method for estimating the six Degrees of Freedom (6DoF) pose of texture-less objects from a monocular image by using edge information. The deep learning-based pose estimation method needs a large dataset containing pairs of an image and ground truth pose of objects. To alleviat...

Descripción completa

Detalles Bibliográficos
Autores principales: Moteki, Atsunori, Saito, Hideo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9781246/
https://www.ncbi.nlm.nih.gov/pubmed/36559978
http://dx.doi.org/10.3390/s22249610
_version_ 1784857026926477312
author Moteki, Atsunori
Saito, Hideo
author_facet Moteki, Atsunori
Saito, Hideo
author_sort Moteki, Atsunori
collection PubMed
description This paper presents a method for estimating the six Degrees of Freedom (6DoF) pose of texture-less objects from a monocular image by using edge information. The deep learning-based pose estimation method needs a large dataset containing pairs of an image and ground truth pose of objects. To alleviate the cost of collecting a dataset, we focus on the method using a dataset made by computer graphics (CG). This simulation-based method prepares a thousand images by rendering the computer-aided design (CAD) data of the object and trains a deep-learning model. As an inference stage, a monocular RGB image is entered into the model, and the object’s pose is estimated. The representative simulation-based method, Pose Interpreter Networks, uses silhouette images as the input, thereby enabling common feature (contour) extraction from RGB and CG images. However, estimating rotation parameters is less accurate. To overcome this problem, we propose a method to use edge information extracted from the object’s ridgelines for training the deep learning model. Since edge distribution changes largely according to the pose, the estimation of rotation parameters becomes more robust. Through an experiment with simulation data, we quantitatively proved the accuracy improvement compared to the previous method (error rate decreases at a certain condition are translation 22.9% and rotation: 43.4%). Moreover, through an experiment with physical data, we clarified the issues of this method and proposed an effective solution by fine-tuning (error rate decrease at a certain condition are translation 20.1% and rotation 57.7%).
format Online
Article
Text
id pubmed-9781246
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-97812462022-12-24 Object Pose Estimation Using Edge Images Synthesized from Shape Information Moteki, Atsunori Saito, Hideo Sensors (Basel) Article This paper presents a method for estimating the six Degrees of Freedom (6DoF) pose of texture-less objects from a monocular image by using edge information. The deep learning-based pose estimation method needs a large dataset containing pairs of an image and ground truth pose of objects. To alleviate the cost of collecting a dataset, we focus on the method using a dataset made by computer graphics (CG). This simulation-based method prepares a thousand images by rendering the computer-aided design (CAD) data of the object and trains a deep-learning model. As an inference stage, a monocular RGB image is entered into the model, and the object’s pose is estimated. The representative simulation-based method, Pose Interpreter Networks, uses silhouette images as the input, thereby enabling common feature (contour) extraction from RGB and CG images. However, estimating rotation parameters is less accurate. To overcome this problem, we propose a method to use edge information extracted from the object’s ridgelines for training the deep learning model. Since edge distribution changes largely according to the pose, the estimation of rotation parameters becomes more robust. Through an experiment with simulation data, we quantitatively proved the accuracy improvement compared to the previous method (error rate decreases at a certain condition are translation 22.9% and rotation: 43.4%). Moreover, through an experiment with physical data, we clarified the issues of this method and proposed an effective solution by fine-tuning (error rate decrease at a certain condition are translation 20.1% and rotation 57.7%). MDPI 2022-12-08 /pmc/articles/PMC9781246/ /pubmed/36559978 http://dx.doi.org/10.3390/s22249610 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Moteki, Atsunori
Saito, Hideo
Object Pose Estimation Using Edge Images Synthesized from Shape Information
title Object Pose Estimation Using Edge Images Synthesized from Shape Information
title_full Object Pose Estimation Using Edge Images Synthesized from Shape Information
title_fullStr Object Pose Estimation Using Edge Images Synthesized from Shape Information
title_full_unstemmed Object Pose Estimation Using Edge Images Synthesized from Shape Information
title_short Object Pose Estimation Using Edge Images Synthesized from Shape Information
title_sort object pose estimation using edge images synthesized from shape information
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9781246/
https://www.ncbi.nlm.nih.gov/pubmed/36559978
http://dx.doi.org/10.3390/s22249610
work_keys_str_mv AT motekiatsunori objectposeestimationusingedgeimagessynthesizedfromshapeinformation
AT saitohideo objectposeestimationusingedgeimagessynthesizedfromshapeinformation