Cargando…

Autonomous Fingerprinting and Large Experimental Data Set for Visible Light Positioning

This paper presents an autonomous method of collecting data for Visible Light Positioning (VLP) and a comprehensive investigation of VLP using a large set of experimental data. Received Signal Strength (RSS) data are efficiently collected using a novel method that utilizes consumer grade Virtual Rea...

Descripción completa

Detalles Bibliográficos
Autores principales: Glass, Tyrel, Alam, Fakhrul, Legg, Mathew, Noble, Frazer
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8125820/
https://www.ncbi.nlm.nih.gov/pubmed/34066704
http://dx.doi.org/10.3390/s21093256
_version_ 1783693613515931648
author Glass, Tyrel
Alam, Fakhrul
Legg, Mathew
Noble, Frazer
author_facet Glass, Tyrel
Alam, Fakhrul
Legg, Mathew
Noble, Frazer
author_sort Glass, Tyrel
collection PubMed
description This paper presents an autonomous method of collecting data for Visible Light Positioning (VLP) and a comprehensive investigation of VLP using a large set of experimental data. Received Signal Strength (RSS) data are efficiently collected using a novel method that utilizes consumer grade Virtual Reality (VR) tracking for accurate ground truth recording. An investigation into the accuracy of the ground truth system showed median and 90th percentile errors of 4.24 and 7.35 mm, respectively. Co-locating a VR tracker with a photodiode-equipped VLP receiver on a mobile robotic platform allows fingerprinting on a scale and accuracy that has not been possible with traditional manual collection methods. RSS data at 7344 locations within a 6.3 × 6.9 m test space fitted with 11 VLP luminaires is collected and has been made available for researchers. The quality and the volume of the data allow for a robust study of Machine Learning (ML)- and channel model-based positioning utilizing visible light. Among the ML-based techniques, ridge regression is found to be the most accurate, outperforming Weighted k Nearest Neighbor, Multilayer Perceptron, and random forest, among others. Model-based positioning is more accurate than ML techniques when a small data set is available for calibration and training. However, if a large data set is available for training, ML-based positioning outperforms its model-based counterparts in terms of localization accuracy.
format Online
Article
Text
id pubmed-8125820
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-81258202021-05-17 Autonomous Fingerprinting and Large Experimental Data Set for Visible Light Positioning Glass, Tyrel Alam, Fakhrul Legg, Mathew Noble, Frazer Sensors (Basel) Article This paper presents an autonomous method of collecting data for Visible Light Positioning (VLP) and a comprehensive investigation of VLP using a large set of experimental data. Received Signal Strength (RSS) data are efficiently collected using a novel method that utilizes consumer grade Virtual Reality (VR) tracking for accurate ground truth recording. An investigation into the accuracy of the ground truth system showed median and 90th percentile errors of 4.24 and 7.35 mm, respectively. Co-locating a VR tracker with a photodiode-equipped VLP receiver on a mobile robotic platform allows fingerprinting on a scale and accuracy that has not been possible with traditional manual collection methods. RSS data at 7344 locations within a 6.3 × 6.9 m test space fitted with 11 VLP luminaires is collected and has been made available for researchers. The quality and the volume of the data allow for a robust study of Machine Learning (ML)- and channel model-based positioning utilizing visible light. Among the ML-based techniques, ridge regression is found to be the most accurate, outperforming Weighted k Nearest Neighbor, Multilayer Perceptron, and random forest, among others. Model-based positioning is more accurate than ML techniques when a small data set is available for calibration and training. However, if a large data set is available for training, ML-based positioning outperforms its model-based counterparts in terms of localization accuracy. MDPI 2021-05-08 /pmc/articles/PMC8125820/ /pubmed/34066704 http://dx.doi.org/10.3390/s21093256 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Glass, Tyrel
Alam, Fakhrul
Legg, Mathew
Noble, Frazer
Autonomous Fingerprinting and Large Experimental Data Set for Visible Light Positioning
title Autonomous Fingerprinting and Large Experimental Data Set for Visible Light Positioning
title_full Autonomous Fingerprinting and Large Experimental Data Set for Visible Light Positioning
title_fullStr Autonomous Fingerprinting and Large Experimental Data Set for Visible Light Positioning
title_full_unstemmed Autonomous Fingerprinting and Large Experimental Data Set for Visible Light Positioning
title_short Autonomous Fingerprinting and Large Experimental Data Set for Visible Light Positioning
title_sort autonomous fingerprinting and large experimental data set for visible light positioning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8125820/
https://www.ncbi.nlm.nih.gov/pubmed/34066704
http://dx.doi.org/10.3390/s21093256
work_keys_str_mv AT glasstyrel autonomousfingerprintingandlargeexperimentaldatasetforvisiblelightpositioning
AT alamfakhrul autonomousfingerprintingandlargeexperimentaldatasetforvisiblelightpositioning
AT leggmathew autonomousfingerprintingandlargeexperimentaldatasetforvisiblelightpositioning
AT noblefrazer autonomousfingerprintingandlargeexperimentaldatasetforvisiblelightpositioning