Cargando…

3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data

Local image features are invariant to in-plane rotations and robust to minor viewpoint changes. However, the current detectors and descriptors for local image features fail to accommodate out-of-plane rotations larger than 25°–30°. Invariance to such viewpoint changes is essential for numerous appli...

Descripción completa

Detalles Bibliográficos
Autores principales: Sahloul, Hamdi, Shirafuji, Shouhei, Ota, Jun
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6359040/
https://www.ncbi.nlm.nih.gov/pubmed/30642092
http://dx.doi.org/10.3390/s19020291
_version_ 1783392135389642752
author Sahloul, Hamdi
Shirafuji, Shouhei
Ota, Jun
author_facet Sahloul, Hamdi
Shirafuji, Shouhei
Ota, Jun
author_sort Sahloul, Hamdi
collection PubMed
description Local image features are invariant to in-plane rotations and robust to minor viewpoint changes. However, the current detectors and descriptors for local image features fail to accommodate out-of-plane rotations larger than 25°–30°. Invariance to such viewpoint changes is essential for numerous applications, including wide baseline matching, 6D pose estimation, and object reconstruction. In this study, we present a general embedding that wraps a detector/descriptor pair in order to increase viewpoint invariance by exploiting input depth maps. The proposed embedding locates smooth surfaces within the input RGB-D images and projects them into a viewpoint invariant representation, enabling the detection and description of more viewpoint invariant features. Our embedding can be utilized with different combinations of descriptor/detector pairs, according to the desired application. Using synthetic and real-world objects, we evaluated the viewpoint invariance of various detectors and descriptors, for both standalone and embedded approaches. While standalone local image features fail to accommodate average viewpoint changes beyond 33.3°, our proposed embedding boosted the viewpoint invariance to different levels, depending on the scene geometry. Objects with distinct surface discontinuities were on average invariant up to 52.8°, and the overall average for all evaluated datasets was 45.4°. Similarly, out of a total of 140 combinations involving 20 local image features and various objects with distinct surface discontinuities, only a single standalone local image feature exceeded the goal of 60° viewpoint difference in just two combinations, as compared with 19 different local image features succeeding in 73 combinations when wrapped in the proposed embedding. Furthermore, the proposed approach operates robustly in the presence of input depth noise, even that of low-cost commodity depth sensors, and well beyond.
format Online
Article
Text
id pubmed-6359040
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-63590402019-02-06 3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data Sahloul, Hamdi Shirafuji, Shouhei Ota, Jun Sensors (Basel) Article Local image features are invariant to in-plane rotations and robust to minor viewpoint changes. However, the current detectors and descriptors for local image features fail to accommodate out-of-plane rotations larger than 25°–30°. Invariance to such viewpoint changes is essential for numerous applications, including wide baseline matching, 6D pose estimation, and object reconstruction. In this study, we present a general embedding that wraps a detector/descriptor pair in order to increase viewpoint invariance by exploiting input depth maps. The proposed embedding locates smooth surfaces within the input RGB-D images and projects them into a viewpoint invariant representation, enabling the detection and description of more viewpoint invariant features. Our embedding can be utilized with different combinations of descriptor/detector pairs, according to the desired application. Using synthetic and real-world objects, we evaluated the viewpoint invariance of various detectors and descriptors, for both standalone and embedded approaches. While standalone local image features fail to accommodate average viewpoint changes beyond 33.3°, our proposed embedding boosted the viewpoint invariance to different levels, depending on the scene geometry. Objects with distinct surface discontinuities were on average invariant up to 52.8°, and the overall average for all evaluated datasets was 45.4°. Similarly, out of a total of 140 combinations involving 20 local image features and various objects with distinct surface discontinuities, only a single standalone local image feature exceeded the goal of 60° viewpoint difference in just two combinations, as compared with 19 different local image features succeeding in 73 combinations when wrapped in the proposed embedding. Furthermore, the proposed approach operates robustly in the presence of input depth noise, even that of low-cost commodity depth sensors, and well beyond. MDPI 2019-01-12 /pmc/articles/PMC6359040/ /pubmed/30642092 http://dx.doi.org/10.3390/s19020291 Text en © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Sahloul, Hamdi
Shirafuji, Shouhei
Ota, Jun
3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data
title 3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data
title_full 3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data
title_fullStr 3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data
title_full_unstemmed 3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data
title_short 3D Affine: An Embedding of Local Image Features for Viewpoint Invariance Using RGB-D Sensor Data
title_sort 3d affine: an embedding of local image features for viewpoint invariance using rgb-d sensor data
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6359040/
https://www.ncbi.nlm.nih.gov/pubmed/30642092
http://dx.doi.org/10.3390/s19020291
work_keys_str_mv AT sahloulhamdi 3daffineanembeddingoflocalimagefeaturesforviewpointinvarianceusingrgbdsensordata
AT shirafujishouhei 3daffineanembeddingoflocalimagefeaturesforviewpointinvarianceusingrgbdsensordata
AT otajun 3daffineanembeddingoflocalimagefeaturesforviewpointinvarianceusingrgbdsensordata