Cargando…

Self-Supervised Robust Feature Matching Pipeline for Teach and Repeat Navigation

The performance of deep neural networks and the low costs of computational hardware has made computer vision a popular choice in many robotic systems. An attractive feature of deep-learned methods is their ability to cope with appearance changes caused by day–night cycles and seasonal variations. Ho...

Descripción completa

Detalles Bibliográficos
Autores principales: Rouček, Tomáš, Amjadi, Arash Sadeghi, Rozsypálek, Zdeněk, Broughton, George, Blaha, Jan, Kusumam, Keerthy, Krajník, Tomáš
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9032253/
https://www.ncbi.nlm.nih.gov/pubmed/35458823
http://dx.doi.org/10.3390/s22082836
_version_ 1784692596190216192
author Rouček, Tomáš
Amjadi, Arash Sadeghi
Rozsypálek, Zdeněk
Broughton, George
Blaha, Jan
Kusumam, Keerthy
Krajník, Tomáš
author_facet Rouček, Tomáš
Amjadi, Arash Sadeghi
Rozsypálek, Zdeněk
Broughton, George
Blaha, Jan
Kusumam, Keerthy
Krajník, Tomáš
author_sort Rouček, Tomáš
collection PubMed
description The performance of deep neural networks and the low costs of computational hardware has made computer vision a popular choice in many robotic systems. An attractive feature of deep-learned methods is their ability to cope with appearance changes caused by day–night cycles and seasonal variations. However, deep learning of neural networks typically relies on large numbers of hand-annotated images, which requires significant effort for data collection and annotation. We present a method that allows autonomous, self-supervised training of a neural network in visual teach-and-repeat (VT&R) tasks, where a mobile robot has to traverse a previously taught path repeatedly. Our method is based on a fusion of two image registration schemes: one based on a Siamese neural network and another on point-feature matching. As the robot traverses the taught paths, it uses the results of feature-based matching to train the neural network, which, in turn, provides coarse registration estimates to the feature matcher. We show that as the neural network gets trained, the accuracy and robustness of the navigation increases, making the robot capable of dealing with significant changes in the environment. This method can significantly reduce the data annotation efforts when designing new robotic systems or introducing robots into new environments. Moreover, the method provides annotated datasets that can be deployed in other navigation systems. To promote the reproducibility of the research presented herein, we provide our datasets, codes and trained models online.
format Online
Article
Text
id pubmed-9032253
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-90322532022-04-23 Self-Supervised Robust Feature Matching Pipeline for Teach and Repeat Navigation Rouček, Tomáš Amjadi, Arash Sadeghi Rozsypálek, Zdeněk Broughton, George Blaha, Jan Kusumam, Keerthy Krajník, Tomáš Sensors (Basel) Article The performance of deep neural networks and the low costs of computational hardware has made computer vision a popular choice in many robotic systems. An attractive feature of deep-learned methods is their ability to cope with appearance changes caused by day–night cycles and seasonal variations. However, deep learning of neural networks typically relies on large numbers of hand-annotated images, which requires significant effort for data collection and annotation. We present a method that allows autonomous, self-supervised training of a neural network in visual teach-and-repeat (VT&R) tasks, where a mobile robot has to traverse a previously taught path repeatedly. Our method is based on a fusion of two image registration schemes: one based on a Siamese neural network and another on point-feature matching. As the robot traverses the taught paths, it uses the results of feature-based matching to train the neural network, which, in turn, provides coarse registration estimates to the feature matcher. We show that as the neural network gets trained, the accuracy and robustness of the navigation increases, making the robot capable of dealing with significant changes in the environment. This method can significantly reduce the data annotation efforts when designing new robotic systems or introducing robots into new environments. Moreover, the method provides annotated datasets that can be deployed in other navigation systems. To promote the reproducibility of the research presented herein, we provide our datasets, codes and trained models online. MDPI 2022-04-07 /pmc/articles/PMC9032253/ /pubmed/35458823 http://dx.doi.org/10.3390/s22082836 Text en © 2022 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Rouček, Tomáš
Amjadi, Arash Sadeghi
Rozsypálek, Zdeněk
Broughton, George
Blaha, Jan
Kusumam, Keerthy
Krajník, Tomáš
Self-Supervised Robust Feature Matching Pipeline for Teach and Repeat Navigation
title Self-Supervised Robust Feature Matching Pipeline for Teach and Repeat Navigation
title_full Self-Supervised Robust Feature Matching Pipeline for Teach and Repeat Navigation
title_fullStr Self-Supervised Robust Feature Matching Pipeline for Teach and Repeat Navigation
title_full_unstemmed Self-Supervised Robust Feature Matching Pipeline for Teach and Repeat Navigation
title_short Self-Supervised Robust Feature Matching Pipeline for Teach and Repeat Navigation
title_sort self-supervised robust feature matching pipeline for teach and repeat navigation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9032253/
https://www.ncbi.nlm.nih.gov/pubmed/35458823
http://dx.doi.org/10.3390/s22082836
work_keys_str_mv AT roucektomas selfsupervisedrobustfeaturematchingpipelineforteachandrepeatnavigation
AT amjadiarashsadeghi selfsupervisedrobustfeaturematchingpipelineforteachandrepeatnavigation
AT rozsypalekzdenek selfsupervisedrobustfeaturematchingpipelineforteachandrepeatnavigation
AT broughtongeorge selfsupervisedrobustfeaturematchingpipelineforteachandrepeatnavigation
AT blahajan selfsupervisedrobustfeaturematchingpipelineforteachandrepeatnavigation
AT kusumamkeerthy selfsupervisedrobustfeaturematchingpipelineforteachandrepeatnavigation
AT krajniktomas selfsupervisedrobustfeaturematchingpipelineforteachandrepeatnavigation