Cargando…
3DPointCaps++: Learning 3D Representations with Capsule Networks
We present 3DPointCaps++ for learning robust, flexible and generalizable 3D object representations without requiring heavy annotation efforts or supervision. Unlike conventional 3D generative models, our algorithm aims for building a structured latent space where certain factors of shape variations,...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer US
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9362689/ https://www.ncbi.nlm.nih.gov/pubmed/35968252 http://dx.doi.org/10.1007/s11263-022-01632-6 |
_version_ | 1784764768176832512 |
---|---|
author | Zhao, Yongheng Fang, Guangchi Guo, Yulan Guibas, Leonidas Tombari, Federico Birdal, Tolga |
author_facet | Zhao, Yongheng Fang, Guangchi Guo, Yulan Guibas, Leonidas Tombari, Federico Birdal, Tolga |
author_sort | Zhao, Yongheng |
collection | PubMed |
description | We present 3DPointCaps++ for learning robust, flexible and generalizable 3D object representations without requiring heavy annotation efforts or supervision. Unlike conventional 3D generative models, our algorithm aims for building a structured latent space where certain factors of shape variations, such as object parts, can be disentangled into independent sub-spaces. Our novel decoder then acts on these individual latent sub-spaces (i.e. capsules) using deconvolution operators to reconstruct 3D points in a self-supervised manner. We further introduce a cluster loss ensuring that the points reconstructed by a single capsule remain local and do not spread across the object uncontrollably. These contributions allow our network to tackle the challenging tasks of part segmentation, part interpolation/replacement as well as correspondence estimation across rigid / non-rigid shape, and across / within category. Our extensive evaluations on ShapeNet objects and human scans demonstrate that our network can learn generic representations that are robust and useful in many applications. |
format | Online Article Text |
id | pubmed-9362689 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Springer US |
record_format | MEDLINE/PubMed |
spelling | pubmed-93626892022-08-10 3DPointCaps++: Learning 3D Representations with Capsule Networks Zhao, Yongheng Fang, Guangchi Guo, Yulan Guibas, Leonidas Tombari, Federico Birdal, Tolga Int J Comput Vis Article We present 3DPointCaps++ for learning robust, flexible and generalizable 3D object representations without requiring heavy annotation efforts or supervision. Unlike conventional 3D generative models, our algorithm aims for building a structured latent space where certain factors of shape variations, such as object parts, can be disentangled into independent sub-spaces. Our novel decoder then acts on these individual latent sub-spaces (i.e. capsules) using deconvolution operators to reconstruct 3D points in a self-supervised manner. We further introduce a cluster loss ensuring that the points reconstructed by a single capsule remain local and do not spread across the object uncontrollably. These contributions allow our network to tackle the challenging tasks of part segmentation, part interpolation/replacement as well as correspondence estimation across rigid / non-rigid shape, and across / within category. Our extensive evaluations on ShapeNet objects and human scans demonstrate that our network can learn generic representations that are robust and useful in many applications. Springer US 2022-07-30 2022 /pmc/articles/PMC9362689/ /pubmed/35968252 http://dx.doi.org/10.1007/s11263-022-01632-6 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Zhao, Yongheng Fang, Guangchi Guo, Yulan Guibas, Leonidas Tombari, Federico Birdal, Tolga 3DPointCaps++: Learning 3D Representations with Capsule Networks |
title | 3DPointCaps++: Learning 3D Representations with Capsule Networks |
title_full | 3DPointCaps++: Learning 3D Representations with Capsule Networks |
title_fullStr | 3DPointCaps++: Learning 3D Representations with Capsule Networks |
title_full_unstemmed | 3DPointCaps++: Learning 3D Representations with Capsule Networks |
title_short | 3DPointCaps++: Learning 3D Representations with Capsule Networks |
title_sort | 3dpointcaps++: learning 3d representations with capsule networks |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9362689/ https://www.ncbi.nlm.nih.gov/pubmed/35968252 http://dx.doi.org/10.1007/s11263-022-01632-6 |
work_keys_str_mv | AT zhaoyongheng 3dpointcapslearning3drepresentationswithcapsulenetworks AT fangguangchi 3dpointcapslearning3drepresentationswithcapsulenetworks AT guoyulan 3dpointcapslearning3drepresentationswithcapsulenetworks AT guibasleonidas 3dpointcapslearning3drepresentationswithcapsulenetworks AT tombarifederico 3dpointcapslearning3drepresentationswithcapsulenetworks AT birdaltolga 3dpointcapslearning3drepresentationswithcapsulenetworks |