Cargando…
Visual state estimation in unseen environments through domain adaptation and metric learning
In robotics, deep learning models are used in many visual perception applications, including the tracking, detection and pose estimation of robotic manipulators. The state of the art methods however are conditioned on the availability of annotated training data, which may in practice be costly or ev...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9437705/ https://www.ncbi.nlm.nih.gov/pubmed/36059568 http://dx.doi.org/10.3389/frobt.2022.833173 |
_version_ | 1784781677432668160 |
---|---|
author | Güler, Püren Stork, Johannes A. Stoyanov, Todor |
author_facet | Güler, Püren Stork, Johannes A. Stoyanov, Todor |
author_sort | Güler, Püren |
collection | PubMed |
description | In robotics, deep learning models are used in many visual perception applications, including the tracking, detection and pose estimation of robotic manipulators. The state of the art methods however are conditioned on the availability of annotated training data, which may in practice be costly or even impossible to collect. Domain augmentation is one popular method to improve generalization to out-of-domain data by extending the training data set with predefined sources of variation, unrelated to the primary task. While this typically results in better performance on the target domain, it is not always clear that the trained models are capable to accurately separate the signals relevant to solving the task (e.g., appearance of an object of interest) from those associated with differences between the domains (e.g., lighting conditions). In this work we propose to improve the generalization capabilities of models trained with domain augmentation by formulating a secondary structured metric-space learning objective. We concentrate on one particularly challenging domain transfer task—visual state estimation for an articulated underground mining machine—and demonstrate the benefits of imposing structure on the encoding space. Our results indicate that the proposed method has the potential to transfer feature embeddings learned on the source domain, through a suitably designed augmentation procedure, and on to an unseen target domain. |
format | Online Article Text |
id | pubmed-9437705 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-94377052022-09-03 Visual state estimation in unseen environments through domain adaptation and metric learning Güler, Püren Stork, Johannes A. Stoyanov, Todor Front Robot AI Robotics and AI In robotics, deep learning models are used in many visual perception applications, including the tracking, detection and pose estimation of robotic manipulators. The state of the art methods however are conditioned on the availability of annotated training data, which may in practice be costly or even impossible to collect. Domain augmentation is one popular method to improve generalization to out-of-domain data by extending the training data set with predefined sources of variation, unrelated to the primary task. While this typically results in better performance on the target domain, it is not always clear that the trained models are capable to accurately separate the signals relevant to solving the task (e.g., appearance of an object of interest) from those associated with differences between the domains (e.g., lighting conditions). In this work we propose to improve the generalization capabilities of models trained with domain augmentation by formulating a secondary structured metric-space learning objective. We concentrate on one particularly challenging domain transfer task—visual state estimation for an articulated underground mining machine—and demonstrate the benefits of imposing structure on the encoding space. Our results indicate that the proposed method has the potential to transfer feature embeddings learned on the source domain, through a suitably designed augmentation procedure, and on to an unseen target domain. Frontiers Media S.A. 2022-08-19 /pmc/articles/PMC9437705/ /pubmed/36059568 http://dx.doi.org/10.3389/frobt.2022.833173 Text en Copyright © 2022 Güler, Stork and Stoyanov. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Robotics and AI Güler, Püren Stork, Johannes A. Stoyanov, Todor Visual state estimation in unseen environments through domain adaptation and metric learning |
title | Visual state estimation in unseen environments through domain adaptation and metric learning |
title_full | Visual state estimation in unseen environments through domain adaptation and metric learning |
title_fullStr | Visual state estimation in unseen environments through domain adaptation and metric learning |
title_full_unstemmed | Visual state estimation in unseen environments through domain adaptation and metric learning |
title_short | Visual state estimation in unseen environments through domain adaptation and metric learning |
title_sort | visual state estimation in unseen environments through domain adaptation and metric learning |
topic | Robotics and AI |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9437705/ https://www.ncbi.nlm.nih.gov/pubmed/36059568 http://dx.doi.org/10.3389/frobt.2022.833173 |
work_keys_str_mv | AT gulerpuren visualstateestimationinunseenenvironmentsthroughdomainadaptationandmetriclearning AT storkjohannesa visualstateestimationinunseenenvironmentsthroughdomainadaptationandmetriclearning AT stoyanovtodor visualstateestimationinunseenenvironmentsthroughdomainadaptationandmetriclearning |