Cargando…

Visual state estimation in unseen environments through domain adaptation and metric learning

In robotics, deep learning models are used in many visual perception applications, including the tracking, detection and pose estimation of robotic manipulators. The state of the art methods however are conditioned on the availability of annotated training data, which may in practice be costly or ev...

Descripción completa

Detalles Bibliográficos
Autores principales: Güler, Püren, Stork, Johannes A., Stoyanov, Todor
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9437705/
https://www.ncbi.nlm.nih.gov/pubmed/36059568
http://dx.doi.org/10.3389/frobt.2022.833173
_version_ 1784781677432668160
author Güler, Püren
Stork, Johannes A.
Stoyanov, Todor
author_facet Güler, Püren
Stork, Johannes A.
Stoyanov, Todor
author_sort Güler, Püren
collection PubMed
description In robotics, deep learning models are used in many visual perception applications, including the tracking, detection and pose estimation of robotic manipulators. The state of the art methods however are conditioned on the availability of annotated training data, which may in practice be costly or even impossible to collect. Domain augmentation is one popular method to improve generalization to out-of-domain data by extending the training data set with predefined sources of variation, unrelated to the primary task. While this typically results in better performance on the target domain, it is not always clear that the trained models are capable to accurately separate the signals relevant to solving the task (e.g., appearance of an object of interest) from those associated with differences between the domains (e.g., lighting conditions). In this work we propose to improve the generalization capabilities of models trained with domain augmentation by formulating a secondary structured metric-space learning objective. We concentrate on one particularly challenging domain transfer task—visual state estimation for an articulated underground mining machine—and demonstrate the benefits of imposing structure on the encoding space. Our results indicate that the proposed method has the potential to transfer feature embeddings learned on the source domain, through a suitably designed augmentation procedure, and on to an unseen target domain.
format Online
Article
Text
id pubmed-9437705
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-94377052022-09-03 Visual state estimation in unseen environments through domain adaptation and metric learning Güler, Püren Stork, Johannes A. Stoyanov, Todor Front Robot AI Robotics and AI In robotics, deep learning models are used in many visual perception applications, including the tracking, detection and pose estimation of robotic manipulators. The state of the art methods however are conditioned on the availability of annotated training data, which may in practice be costly or even impossible to collect. Domain augmentation is one popular method to improve generalization to out-of-domain data by extending the training data set with predefined sources of variation, unrelated to the primary task. While this typically results in better performance on the target domain, it is not always clear that the trained models are capable to accurately separate the signals relevant to solving the task (e.g., appearance of an object of interest) from those associated with differences between the domains (e.g., lighting conditions). In this work we propose to improve the generalization capabilities of models trained with domain augmentation by formulating a secondary structured metric-space learning objective. We concentrate on one particularly challenging domain transfer task—visual state estimation for an articulated underground mining machine—and demonstrate the benefits of imposing structure on the encoding space. Our results indicate that the proposed method has the potential to transfer feature embeddings learned on the source domain, through a suitably designed augmentation procedure, and on to an unseen target domain. Frontiers Media S.A. 2022-08-19 /pmc/articles/PMC9437705/ /pubmed/36059568 http://dx.doi.org/10.3389/frobt.2022.833173 Text en Copyright © 2022 Güler, Stork and Stoyanov. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Robotics and AI
Güler, Püren
Stork, Johannes A.
Stoyanov, Todor
Visual state estimation in unseen environments through domain adaptation and metric learning
title Visual state estimation in unseen environments through domain adaptation and metric learning
title_full Visual state estimation in unseen environments through domain adaptation and metric learning
title_fullStr Visual state estimation in unseen environments through domain adaptation and metric learning
title_full_unstemmed Visual state estimation in unseen environments through domain adaptation and metric learning
title_short Visual state estimation in unseen environments through domain adaptation and metric learning
title_sort visual state estimation in unseen environments through domain adaptation and metric learning
topic Robotics and AI
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9437705/
https://www.ncbi.nlm.nih.gov/pubmed/36059568
http://dx.doi.org/10.3389/frobt.2022.833173
work_keys_str_mv AT gulerpuren visualstateestimationinunseenenvironmentsthroughdomainadaptationandmetriclearning
AT storkjohannesa visualstateestimationinunseenenvironmentsthroughdomainadaptationandmetriclearning
AT stoyanovtodor visualstateestimationinunseenenvironmentsthroughdomainadaptationandmetriclearning