Cargando…

Sensor Data Fusion for a Mobile Robot Using Neural Networks

Mobile robots must be capable to obtain an accurate map of their surroundings to move within it. To detect different materials that might be undetectable to one sensor but not others it is necessary to construct at least a two-sensor fusion scheme. With this, it is possible to generate a 2D occupanc...

Descripción completa

Detalles Bibliográficos
Autores principales: Barreto-Cubero, Andres J., Gómez-Espinosa, Alfonso, Escobedo Cabello, Jesús Arturo, Cuan-Urquizo, Enrique, Cruz-Ramírez, Sergio R.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8749872/
https://www.ncbi.nlm.nih.gov/pubmed/35009849
http://dx.doi.org/10.3390/s22010305
_version_ 1784631333437308928
author Barreto-Cubero, Andres J.
Gómez-Espinosa, Alfonso
Escobedo Cabello, Jesús Arturo
Cuan-Urquizo, Enrique
Cruz-Ramírez, Sergio R.
author_facet Barreto-Cubero, Andres J.
Gómez-Espinosa, Alfonso
Escobedo Cabello, Jesús Arturo
Cuan-Urquizo, Enrique
Cruz-Ramírez, Sergio R.
author_sort Barreto-Cubero, Andres J.
collection PubMed
description Mobile robots must be capable to obtain an accurate map of their surroundings to move within it. To detect different materials that might be undetectable to one sensor but not others it is necessary to construct at least a two-sensor fusion scheme. With this, it is possible to generate a 2D occupancy map in which glass obstacles are identified. An artificial neural network is used to fuse data from a tri-sensor (RealSense Stereo camera, 2D [Formula: see text] LiDAR, and Ultrasonic Sensors) setup capable of detecting glass and other materials typically found in indoor environments that may or may not be visible to traditional 2D LiDAR sensors, hence the expression improved LiDAR. A preprocessing scheme is implemented to filter all the outliers, project a 3D pointcloud to a 2D plane and adjust distance data. With a Neural Network as a data fusion algorithm, we integrate all the information into a single, more accurate distance-to-obstacle reading to finally generate a 2D Occupancy Grid Map (OGM) that considers all sensors information. The Robotis Turtlebot3 Waffle Pi robot is used as the experimental platform to conduct experiments given the different fusion strategies. Test results show that with such a fusion algorithm, it is possible to detect glass and other obstacles with an estimated root-mean-square error (RMSE) of 3 cm with multiple fusion strategies.
format Online
Article
Text
id pubmed-8749872
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-87498722022-01-12 Sensor Data Fusion for a Mobile Robot Using Neural Networks Barreto-Cubero, Andres J. Gómez-Espinosa, Alfonso Escobedo Cabello, Jesús Arturo Cuan-Urquizo, Enrique Cruz-Ramírez, Sergio R. Sensors (Basel) Article Mobile robots must be capable to obtain an accurate map of their surroundings to move within it. To detect different materials that might be undetectable to one sensor but not others it is necessary to construct at least a two-sensor fusion scheme. With this, it is possible to generate a 2D occupancy map in which glass obstacles are identified. An artificial neural network is used to fuse data from a tri-sensor (RealSense Stereo camera, 2D [Formula: see text] LiDAR, and Ultrasonic Sensors) setup capable of detecting glass and other materials typically found in indoor environments that may or may not be visible to traditional 2D LiDAR sensors, hence the expression improved LiDAR. A preprocessing scheme is implemented to filter all the outliers, project a 3D pointcloud to a 2D plane and adjust distance data. With a Neural Network as a data fusion algorithm, we integrate all the information into a single, more accurate distance-to-obstacle reading to finally generate a 2D Occupancy Grid Map (OGM) that considers all sensors information. The Robotis Turtlebot3 Waffle Pi robot is used as the experimental platform to conduct experiments given the different fusion strategies. Test results show that with such a fusion algorithm, it is possible to detect glass and other obstacles with an estimated root-mean-square error (RMSE) of 3 cm with multiple fusion strategies. MDPI 2021-12-31 /pmc/articles/PMC8749872/ /pubmed/35009849 http://dx.doi.org/10.3390/s22010305 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Barreto-Cubero, Andres J.
Gómez-Espinosa, Alfonso
Escobedo Cabello, Jesús Arturo
Cuan-Urquizo, Enrique
Cruz-Ramírez, Sergio R.
Sensor Data Fusion for a Mobile Robot Using Neural Networks
title Sensor Data Fusion for a Mobile Robot Using Neural Networks
title_full Sensor Data Fusion for a Mobile Robot Using Neural Networks
title_fullStr Sensor Data Fusion for a Mobile Robot Using Neural Networks
title_full_unstemmed Sensor Data Fusion for a Mobile Robot Using Neural Networks
title_short Sensor Data Fusion for a Mobile Robot Using Neural Networks
title_sort sensor data fusion for a mobile robot using neural networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8749872/
https://www.ncbi.nlm.nih.gov/pubmed/35009849
http://dx.doi.org/10.3390/s22010305
work_keys_str_mv AT barretocuberoandresj sensordatafusionforamobilerobotusingneuralnetworks
AT gomezespinosaalfonso sensordatafusionforamobilerobotusingneuralnetworks
AT escobedocabellojesusarturo sensordatafusionforamobilerobotusingneuralnetworks
AT cuanurquizoenrique sensordatafusionforamobilerobotusingneuralnetworks
AT cruzramirezsergior sensordatafusionforamobilerobotusingneuralnetworks