Cargando…

A Visual and VAE Based Hierarchical Indoor Localization Method

Precise localization and pose estimation in indoor environments are commonly employed in a wide range of applications, including robotics, augmented reality, and navigation and positioning services. Such applications can be solved via visual-based localization using a pre-built 3D model. The increas...

Descripción completa

Detalles Bibliográficos
Autores principales: Jiang, Jie, Zou, Yin, Chen, Lidong, Fang, Yujie
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8153307/
https://www.ncbi.nlm.nih.gov/pubmed/34068306
http://dx.doi.org/10.3390/s21103406
_version_ 1783698771702448128
author Jiang, Jie
Zou, Yin
Chen, Lidong
Fang, Yujie
author_facet Jiang, Jie
Zou, Yin
Chen, Lidong
Fang, Yujie
author_sort Jiang, Jie
collection PubMed
description Precise localization and pose estimation in indoor environments are commonly employed in a wide range of applications, including robotics, augmented reality, and navigation and positioning services. Such applications can be solved via visual-based localization using a pre-built 3D model. The increase in searching space associated with large scenes can be overcome by retrieving images in advance and subsequently estimating the pose. The majority of current deep learning-based image retrieval methods require labeled data, which increase data annotation costs and complicate the acquisition of data. In this paper, we propose an unsupervised hierarchical indoor localization framework that integrates an unsupervised network variational autoencoder (VAE) with a visual-based Structure-from-Motion (SfM) approach in order to extract global and local features. During the localization process, global features are applied for the image retrieval at the level of the scene map in order to obtain candidate images, and are subsequently used to estimate the pose from 2D-3D matches between query and candidate images. RGB images only are used as the input of the proposed localization system, which is both convenient and challenging. Experimental results reveal that the proposed method can localize images within 0.16 m and 4° in the 7-Scenes data sets and 32.8% within 5 m and 20° in the Baidu data set. Furthermore, our proposed method achieves a higher precision compared to advanced methods.
format Online
Article
Text
id pubmed-8153307
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-81533072021-05-27 A Visual and VAE Based Hierarchical Indoor Localization Method Jiang, Jie Zou, Yin Chen, Lidong Fang, Yujie Sensors (Basel) Article Precise localization and pose estimation in indoor environments are commonly employed in a wide range of applications, including robotics, augmented reality, and navigation and positioning services. Such applications can be solved via visual-based localization using a pre-built 3D model. The increase in searching space associated with large scenes can be overcome by retrieving images in advance and subsequently estimating the pose. The majority of current deep learning-based image retrieval methods require labeled data, which increase data annotation costs and complicate the acquisition of data. In this paper, we propose an unsupervised hierarchical indoor localization framework that integrates an unsupervised network variational autoencoder (VAE) with a visual-based Structure-from-Motion (SfM) approach in order to extract global and local features. During the localization process, global features are applied for the image retrieval at the level of the scene map in order to obtain candidate images, and are subsequently used to estimate the pose from 2D-3D matches between query and candidate images. RGB images only are used as the input of the proposed localization system, which is both convenient and challenging. Experimental results reveal that the proposed method can localize images within 0.16 m and 4° in the 7-Scenes data sets and 32.8% within 5 m and 20° in the Baidu data set. Furthermore, our proposed method achieves a higher precision compared to advanced methods. MDPI 2021-05-13 /pmc/articles/PMC8153307/ /pubmed/34068306 http://dx.doi.org/10.3390/s21103406 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Jiang, Jie
Zou, Yin
Chen, Lidong
Fang, Yujie
A Visual and VAE Based Hierarchical Indoor Localization Method
title A Visual and VAE Based Hierarchical Indoor Localization Method
title_full A Visual and VAE Based Hierarchical Indoor Localization Method
title_fullStr A Visual and VAE Based Hierarchical Indoor Localization Method
title_full_unstemmed A Visual and VAE Based Hierarchical Indoor Localization Method
title_short A Visual and VAE Based Hierarchical Indoor Localization Method
title_sort visual and vae based hierarchical indoor localization method
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8153307/
https://www.ncbi.nlm.nih.gov/pubmed/34068306
http://dx.doi.org/10.3390/s21103406
work_keys_str_mv AT jiangjie avisualandvaebasedhierarchicalindoorlocalizationmethod
AT zouyin avisualandvaebasedhierarchicalindoorlocalizationmethod
AT chenlidong avisualandvaebasedhierarchicalindoorlocalizationmethod
AT fangyujie avisualandvaebasedhierarchicalindoorlocalizationmethod
AT jiangjie visualandvaebasedhierarchicalindoorlocalizationmethod
AT zouyin visualandvaebasedhierarchicalindoorlocalizationmethod
AT chenlidong visualandvaebasedhierarchicalindoorlocalizationmethod
AT fangyujie visualandvaebasedhierarchicalindoorlocalizationmethod