Cargando…

OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over

In the context of collaborative robotics, handing over hand-held objects to a robot is a safety-critical task. Therefore, a robust distinction between human hands and presented objects in image data is essential to avoid contact with robotic grippers. To be able to develop machine learning methods f...

Descripción completa

Detalles Bibliográficos
Autores principales: Stephan, Benedict, Köhler, Mona, Müller, Steffen, Zhang, Yan, Gross, Horst-Michael, Notni, Gunther
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10537499/
https://www.ncbi.nlm.nih.gov/pubmed/37765862
http://dx.doi.org/10.3390/s23187807
_version_ 1785113117369303040
author Stephan, Benedict
Köhler, Mona
Müller, Steffen
Zhang, Yan
Gross, Horst-Michael
Notni, Gunther
author_facet Stephan, Benedict
Köhler, Mona
Müller, Steffen
Zhang, Yan
Gross, Horst-Michael
Notni, Gunther
author_sort Stephan, Benedict
collection PubMed
description In the context of collaborative robotics, handing over hand-held objects to a robot is a safety-critical task. Therefore, a robust distinction between human hands and presented objects in image data is essential to avoid contact with robotic grippers. To be able to develop machine learning methods for solving this problem, we created the OHO (Object Hand-Over) dataset of tools and other everyday objects being held by human hands. Our dataset consists of color, depth, and thermal images with the addition of pose and shape information about the objects in a real-world scenario. Although the focus of this paper is on instance segmentation, our dataset also enables training for different tasks such as 3D pose estimation or shape estimation of objects. For the instance segmentation task, we present a pipeline for automated label generation in point clouds, as well as image data. Through baseline experiments, we show that these labels are suitable for training an instance segmentation to distinguish hands from objects on a per-pixel basis. Moreover, we present qualitative results for applying our trained model in a real-world application.
format Online
Article
Text
id pubmed-10537499
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-105374992023-09-29 OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over Stephan, Benedict Köhler, Mona Müller, Steffen Zhang, Yan Gross, Horst-Michael Notni, Gunther Sensors (Basel) Article In the context of collaborative robotics, handing over hand-held objects to a robot is a safety-critical task. Therefore, a robust distinction between human hands and presented objects in image data is essential to avoid contact with robotic grippers. To be able to develop machine learning methods for solving this problem, we created the OHO (Object Hand-Over) dataset of tools and other everyday objects being held by human hands. Our dataset consists of color, depth, and thermal images with the addition of pose and shape information about the objects in a real-world scenario. Although the focus of this paper is on instance segmentation, our dataset also enables training for different tasks such as 3D pose estimation or shape estimation of objects. For the instance segmentation task, we present a pipeline for automated label generation in point clouds, as well as image data. Through baseline experiments, we show that these labels are suitable for training an instance segmentation to distinguish hands from objects on a per-pixel basis. Moreover, we present qualitative results for applying our trained model in a real-world application. MDPI 2023-09-11 /pmc/articles/PMC10537499/ /pubmed/37765862 http://dx.doi.org/10.3390/s23187807 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Stephan, Benedict
Köhler, Mona
Müller, Steffen
Zhang, Yan
Gross, Horst-Michael
Notni, Gunther
OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over
title OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over
title_full OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over
title_fullStr OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over
title_full_unstemmed OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over
title_short OHO: A Multi-Modal, Multi-Purpose Dataset for Human-Robot Object Hand-Over
title_sort oho: a multi-modal, multi-purpose dataset for human-robot object hand-over
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10537499/
https://www.ncbi.nlm.nih.gov/pubmed/37765862
http://dx.doi.org/10.3390/s23187807
work_keys_str_mv AT stephanbenedict ohoamultimodalmultipurposedatasetforhumanrobotobjecthandover
AT kohlermona ohoamultimodalmultipurposedatasetforhumanrobotobjecthandover
AT mullersteffen ohoamultimodalmultipurposedatasetforhumanrobotobjecthandover
AT zhangyan ohoamultimodalmultipurposedatasetforhumanrobotobjecthandover
AT grosshorstmichael ohoamultimodalmultipurposedatasetforhumanrobotobjecthandover
AT notnigunther ohoamultimodalmultipurposedatasetforhumanrobotobjecthandover