Cargando…
A Visual and VAE Based Hierarchical Indoor Localization Method
Precise localization and pose estimation in indoor environments are commonly employed in a wide range of applications, including robotics, augmented reality, and navigation and positioning services. Such applications can be solved via visual-based localization using a pre-built 3D model. The increas...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8153307/ https://www.ncbi.nlm.nih.gov/pubmed/34068306 http://dx.doi.org/10.3390/s21103406 |
_version_ | 1783698771702448128 |
---|---|
author | Jiang, Jie Zou, Yin Chen, Lidong Fang, Yujie |
author_facet | Jiang, Jie Zou, Yin Chen, Lidong Fang, Yujie |
author_sort | Jiang, Jie |
collection | PubMed |
description | Precise localization and pose estimation in indoor environments are commonly employed in a wide range of applications, including robotics, augmented reality, and navigation and positioning services. Such applications can be solved via visual-based localization using a pre-built 3D model. The increase in searching space associated with large scenes can be overcome by retrieving images in advance and subsequently estimating the pose. The majority of current deep learning-based image retrieval methods require labeled data, which increase data annotation costs and complicate the acquisition of data. In this paper, we propose an unsupervised hierarchical indoor localization framework that integrates an unsupervised network variational autoencoder (VAE) with a visual-based Structure-from-Motion (SfM) approach in order to extract global and local features. During the localization process, global features are applied for the image retrieval at the level of the scene map in order to obtain candidate images, and are subsequently used to estimate the pose from 2D-3D matches between query and candidate images. RGB images only are used as the input of the proposed localization system, which is both convenient and challenging. Experimental results reveal that the proposed method can localize images within 0.16 m and 4° in the 7-Scenes data sets and 32.8% within 5 m and 20° in the Baidu data set. Furthermore, our proposed method achieves a higher precision compared to advanced methods. |
format | Online Article Text |
id | pubmed-8153307 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-81533072021-05-27 A Visual and VAE Based Hierarchical Indoor Localization Method Jiang, Jie Zou, Yin Chen, Lidong Fang, Yujie Sensors (Basel) Article Precise localization and pose estimation in indoor environments are commonly employed in a wide range of applications, including robotics, augmented reality, and navigation and positioning services. Such applications can be solved via visual-based localization using a pre-built 3D model. The increase in searching space associated with large scenes can be overcome by retrieving images in advance and subsequently estimating the pose. The majority of current deep learning-based image retrieval methods require labeled data, which increase data annotation costs and complicate the acquisition of data. In this paper, we propose an unsupervised hierarchical indoor localization framework that integrates an unsupervised network variational autoencoder (VAE) with a visual-based Structure-from-Motion (SfM) approach in order to extract global and local features. During the localization process, global features are applied for the image retrieval at the level of the scene map in order to obtain candidate images, and are subsequently used to estimate the pose from 2D-3D matches between query and candidate images. RGB images only are used as the input of the proposed localization system, which is both convenient and challenging. Experimental results reveal that the proposed method can localize images within 0.16 m and 4° in the 7-Scenes data sets and 32.8% within 5 m and 20° in the Baidu data set. Furthermore, our proposed method achieves a higher precision compared to advanced methods. MDPI 2021-05-13 /pmc/articles/PMC8153307/ /pubmed/34068306 http://dx.doi.org/10.3390/s21103406 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Jiang, Jie Zou, Yin Chen, Lidong Fang, Yujie A Visual and VAE Based Hierarchical Indoor Localization Method |
title | A Visual and VAE Based Hierarchical Indoor Localization Method |
title_full | A Visual and VAE Based Hierarchical Indoor Localization Method |
title_fullStr | A Visual and VAE Based Hierarchical Indoor Localization Method |
title_full_unstemmed | A Visual and VAE Based Hierarchical Indoor Localization Method |
title_short | A Visual and VAE Based Hierarchical Indoor Localization Method |
title_sort | visual and vae based hierarchical indoor localization method |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8153307/ https://www.ncbi.nlm.nih.gov/pubmed/34068306 http://dx.doi.org/10.3390/s21103406 |
work_keys_str_mv | AT jiangjie avisualandvaebasedhierarchicalindoorlocalizationmethod AT zouyin avisualandvaebasedhierarchicalindoorlocalizationmethod AT chenlidong avisualandvaebasedhierarchicalindoorlocalizationmethod AT fangyujie avisualandvaebasedhierarchicalindoorlocalizationmethod AT jiangjie visualandvaebasedhierarchicalindoorlocalizationmethod AT zouyin visualandvaebasedhierarchicalindoorlocalizationmethod AT chenlidong visualandvaebasedhierarchicalindoorlocalizationmethod AT fangyujie visualandvaebasedhierarchicalindoorlocalizationmethod |