Cargando…

Semantic Mapping for Autonomous Subsea Intervention

This paper presents a method to build a semantic map to assist an underwater vehicle-manipulator system in performing intervention tasks autonomously in a submerged man-made pipe structure. The method is based on the integration of feature-based simultaneous localization and mapping (SLAM) and 3D ob...

Descripción completa

Detalles Bibliográficos
Autores principales: Vallicrosa, Guillem, Himri, Khadidja, Ridao, Pere, Gracias, Nuno
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8538227/
https://www.ncbi.nlm.nih.gov/pubmed/34695951
http://dx.doi.org/10.3390/s21206740
_version_ 1784588455017185280
author Vallicrosa, Guillem
Himri, Khadidja
Ridao, Pere
Gracias, Nuno
author_facet Vallicrosa, Guillem
Himri, Khadidja
Ridao, Pere
Gracias, Nuno
author_sort Vallicrosa, Guillem
collection PubMed
description This paper presents a method to build a semantic map to assist an underwater vehicle-manipulator system in performing intervention tasks autonomously in a submerged man-made pipe structure. The method is based on the integration of feature-based simultaneous localization and mapping (SLAM) and 3D object recognition using a database of a priori known objects. The robot uses Doppler velocity log (DVL), pressure, and attitude and heading reference system (AHRS) sensors for navigation and is equipped with a laser scanner providing non-coloured 3D point clouds of the inspected structure in real time. The object recognition module recognises the pipes and objects within the scan and passes them to the SLAM, which adds them to the map if not yet observed. Otherwise, it uses them to correct the map and the robot navigation if they were already mapped. The SLAM provides a consistent map and a drift-less navigation. Moreover, it provides a global identifier for every observed object instance and its pipe connectivity. This information is fed back to the object recognition module, where it is used to estimate the object classes using Bayesian techniques over the set of those object classes which are compatible in terms of pipe connectivity. This allows fusing of all the already available object observations to improve recognition. The outcome of the process is a semantic map made of pipes connected through valves, elbows and tees conforming to the real structure. Knowing the class and the position of objects will enable high-level manipulation commands in the near future.
format Online
Article
Text
id pubmed-8538227
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-85382272021-10-24 Semantic Mapping for Autonomous Subsea Intervention Vallicrosa, Guillem Himri, Khadidja Ridao, Pere Gracias, Nuno Sensors (Basel) Article This paper presents a method to build a semantic map to assist an underwater vehicle-manipulator system in performing intervention tasks autonomously in a submerged man-made pipe structure. The method is based on the integration of feature-based simultaneous localization and mapping (SLAM) and 3D object recognition using a database of a priori known objects. The robot uses Doppler velocity log (DVL), pressure, and attitude and heading reference system (AHRS) sensors for navigation and is equipped with a laser scanner providing non-coloured 3D point clouds of the inspected structure in real time. The object recognition module recognises the pipes and objects within the scan and passes them to the SLAM, which adds them to the map if not yet observed. Otherwise, it uses them to correct the map and the robot navigation if they were already mapped. The SLAM provides a consistent map and a drift-less navigation. Moreover, it provides a global identifier for every observed object instance and its pipe connectivity. This information is fed back to the object recognition module, where it is used to estimate the object classes using Bayesian techniques over the set of those object classes which are compatible in terms of pipe connectivity. This allows fusing of all the already available object observations to improve recognition. The outcome of the process is a semantic map made of pipes connected through valves, elbows and tees conforming to the real structure. Knowing the class and the position of objects will enable high-level manipulation commands in the near future. MDPI 2021-10-11 /pmc/articles/PMC8538227/ /pubmed/34695951 http://dx.doi.org/10.3390/s21206740 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Vallicrosa, Guillem
Himri, Khadidja
Ridao, Pere
Gracias, Nuno
Semantic Mapping for Autonomous Subsea Intervention
title Semantic Mapping for Autonomous Subsea Intervention
title_full Semantic Mapping for Autonomous Subsea Intervention
title_fullStr Semantic Mapping for Autonomous Subsea Intervention
title_full_unstemmed Semantic Mapping for Autonomous Subsea Intervention
title_short Semantic Mapping for Autonomous Subsea Intervention
title_sort semantic mapping for autonomous subsea intervention
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8538227/
https://www.ncbi.nlm.nih.gov/pubmed/34695951
http://dx.doi.org/10.3390/s21206740
work_keys_str_mv AT vallicrosaguillem semanticmappingforautonomoussubseaintervention
AT himrikhadidja semanticmappingforautonomoussubseaintervention
AT ridaopere semanticmappingforautonomoussubseaintervention
AT graciasnuno semanticmappingforautonomoussubseaintervention