Cargando…

3DPointCaps++: Learning 3D Representations with Capsule Networks

We present 3DPointCaps++ for learning robust, flexible and generalizable 3D object representations without requiring heavy annotation efforts or supervision. Unlike conventional 3D generative models, our algorithm aims for building a structured latent space where certain factors of shape variations,...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhao, Yongheng, Fang, Guangchi, Guo, Yulan, Guibas, Leonidas, Tombari, Federico, Birdal, Tolga
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer US 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9362689/
https://www.ncbi.nlm.nih.gov/pubmed/35968252
http://dx.doi.org/10.1007/s11263-022-01632-6
_version_ 1784764768176832512
author Zhao, Yongheng
Fang, Guangchi
Guo, Yulan
Guibas, Leonidas
Tombari, Federico
Birdal, Tolga
author_facet Zhao, Yongheng
Fang, Guangchi
Guo, Yulan
Guibas, Leonidas
Tombari, Federico
Birdal, Tolga
author_sort Zhao, Yongheng
collection PubMed
description We present 3DPointCaps++ for learning robust, flexible and generalizable 3D object representations without requiring heavy annotation efforts or supervision. Unlike conventional 3D generative models, our algorithm aims for building a structured latent space where certain factors of shape variations, such as object parts, can be disentangled into independent sub-spaces. Our novel decoder then acts on these individual latent sub-spaces (i.e. capsules) using deconvolution operators to reconstruct 3D points in a self-supervised manner. We further introduce a cluster loss ensuring that the points reconstructed by a single capsule remain local and do not spread across the object uncontrollably. These contributions allow our network to tackle the challenging tasks of part segmentation, part interpolation/replacement as well as correspondence estimation across rigid / non-rigid shape, and across / within category. Our extensive evaluations on ShapeNet objects and human scans demonstrate that our network can learn generic representations that are robust and useful in many applications.
format Online
Article
Text
id pubmed-9362689
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Springer US
record_format MEDLINE/PubMed
spelling pubmed-93626892022-08-10 3DPointCaps++: Learning 3D Representations with Capsule Networks Zhao, Yongheng Fang, Guangchi Guo, Yulan Guibas, Leonidas Tombari, Federico Birdal, Tolga Int J Comput Vis Article We present 3DPointCaps++ for learning robust, flexible and generalizable 3D object representations without requiring heavy annotation efforts or supervision. Unlike conventional 3D generative models, our algorithm aims for building a structured latent space where certain factors of shape variations, such as object parts, can be disentangled into independent sub-spaces. Our novel decoder then acts on these individual latent sub-spaces (i.e. capsules) using deconvolution operators to reconstruct 3D points in a self-supervised manner. We further introduce a cluster loss ensuring that the points reconstructed by a single capsule remain local and do not spread across the object uncontrollably. These contributions allow our network to tackle the challenging tasks of part segmentation, part interpolation/replacement as well as correspondence estimation across rigid / non-rigid shape, and across / within category. Our extensive evaluations on ShapeNet objects and human scans demonstrate that our network can learn generic representations that are robust and useful in many applications. Springer US 2022-07-30 2022 /pmc/articles/PMC9362689/ /pubmed/35968252 http://dx.doi.org/10.1007/s11263-022-01632-6 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Zhao, Yongheng
Fang, Guangchi
Guo, Yulan
Guibas, Leonidas
Tombari, Federico
Birdal, Tolga
3DPointCaps++: Learning 3D Representations with Capsule Networks
title 3DPointCaps++: Learning 3D Representations with Capsule Networks
title_full 3DPointCaps++: Learning 3D Representations with Capsule Networks
title_fullStr 3DPointCaps++: Learning 3D Representations with Capsule Networks
title_full_unstemmed 3DPointCaps++: Learning 3D Representations with Capsule Networks
title_short 3DPointCaps++: Learning 3D Representations with Capsule Networks
title_sort 3dpointcaps++: learning 3d representations with capsule networks
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9362689/
https://www.ncbi.nlm.nih.gov/pubmed/35968252
http://dx.doi.org/10.1007/s11263-022-01632-6
work_keys_str_mv AT zhaoyongheng 3dpointcapslearning3drepresentationswithcapsulenetworks
AT fangguangchi 3dpointcapslearning3drepresentationswithcapsulenetworks
AT guoyulan 3dpointcapslearning3drepresentationswithcapsulenetworks
AT guibasleonidas 3dpointcapslearning3drepresentationswithcapsulenetworks
AT tombarifederico 3dpointcapslearning3drepresentationswithcapsulenetworks
AT birdaltolga 3dpointcapslearning3drepresentationswithcapsulenetworks