Cargando…

Cutting Pose Prediction from Point Clouds

The challenge of getting machines to understand and interact with natural objects is encountered in important areas such as medicine, agriculture, and, in our case, slaughterhouse automation. Recent breakthroughs have enabled the application of Deep Neural Networks (DNN) directly to point clouds, an...

Descripción completa

Detalles Bibliográficos
Autores principales: Philipsen, Mark P., Moeslund, Thomas B.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7146437/
https://www.ncbi.nlm.nih.gov/pubmed/32168888
http://dx.doi.org/10.3390/s20061563
_version_ 1783520202425630720
author Philipsen, Mark P.
Moeslund, Thomas B.
author_facet Philipsen, Mark P.
Moeslund, Thomas B.
author_sort Philipsen, Mark P.
collection PubMed
description The challenge of getting machines to understand and interact with natural objects is encountered in important areas such as medicine, agriculture, and, in our case, slaughterhouse automation. Recent breakthroughs have enabled the application of Deep Neural Networks (DNN) directly to point clouds, an efficient and natural representation of 3D objects. The potential of these methods has mostly been demonstrated for classification and segmentation tasks involving rigid man-made objects. We present a method, based on the successful PointNet architecture, for learning to regress correct tool placement from human demonstrations, using virtual reality. Our method is applied to a challenging slaughterhouse cutting task, which requires an understanding of the local geometry including the shape, size, and orientation. We propose an intermediate five-Degree of Freedom (DoF) cutting plane representation, a point and a normal vector, which eases the demonstration and learning process. A live experiment is conducted in order to unveil issues and begin to understand the required accuracy. Eleven cuts are rated by an expert, with [Formula: see text] being rated as acceptable. The error on the test set is subsequently reduced through the addition of more training data and improvements to the DNN. The result is a reduction in the average translation from 1.5 cm to 0.8 cm and the orientation error from 4.59° to 4.48°. The method’s generalization capacity is assessed on a similar task from the slaughterhouse and on the very different public LINEMOD dataset for object pose estimation across view points. In both cases, the method shows promising results. Code, datasets, and other materials are available in Supplementary Materials.
format Online
Article
Text
id pubmed-7146437
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-71464372020-04-15 Cutting Pose Prediction from Point Clouds Philipsen, Mark P. Moeslund, Thomas B. Sensors (Basel) Article The challenge of getting machines to understand and interact with natural objects is encountered in important areas such as medicine, agriculture, and, in our case, slaughterhouse automation. Recent breakthroughs have enabled the application of Deep Neural Networks (DNN) directly to point clouds, an efficient and natural representation of 3D objects. The potential of these methods has mostly been demonstrated for classification and segmentation tasks involving rigid man-made objects. We present a method, based on the successful PointNet architecture, for learning to regress correct tool placement from human demonstrations, using virtual reality. Our method is applied to a challenging slaughterhouse cutting task, which requires an understanding of the local geometry including the shape, size, and orientation. We propose an intermediate five-Degree of Freedom (DoF) cutting plane representation, a point and a normal vector, which eases the demonstration and learning process. A live experiment is conducted in order to unveil issues and begin to understand the required accuracy. Eleven cuts are rated by an expert, with [Formula: see text] being rated as acceptable. The error on the test set is subsequently reduced through the addition of more training data and improvements to the DNN. The result is a reduction in the average translation from 1.5 cm to 0.8 cm and the orientation error from 4.59° to 4.48°. The method’s generalization capacity is assessed on a similar task from the slaughterhouse and on the very different public LINEMOD dataset for object pose estimation across view points. In both cases, the method shows promising results. Code, datasets, and other materials are available in Supplementary Materials. MDPI 2020-03-11 /pmc/articles/PMC7146437/ /pubmed/32168888 http://dx.doi.org/10.3390/s20061563 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Philipsen, Mark P.
Moeslund, Thomas B.
Cutting Pose Prediction from Point Clouds
title Cutting Pose Prediction from Point Clouds
title_full Cutting Pose Prediction from Point Clouds
title_fullStr Cutting Pose Prediction from Point Clouds
title_full_unstemmed Cutting Pose Prediction from Point Clouds
title_short Cutting Pose Prediction from Point Clouds
title_sort cutting pose prediction from point clouds
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7146437/
https://www.ncbi.nlm.nih.gov/pubmed/32168888
http://dx.doi.org/10.3390/s20061563
work_keys_str_mv AT philipsenmarkp cuttingposepredictionfrompointclouds
AT moeslundthomasb cuttingposepredictionfrompointclouds