Cargando…
Deep Common Semantic Space Embedding for Sketch-Based 3D Model Retrieval
Sketch-based 3D model retrieval has become an important research topic in many applications, such as computer graphics and computer-aided design. Although sketches and 3D models have huge interdomain visual perception discrepancies, and sketches of the same object have remarkable intradomain visual...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7514853/ https://www.ncbi.nlm.nih.gov/pubmed/33267083 http://dx.doi.org/10.3390/e21040369 |
_version_ | 1783586684564144128 |
---|---|
author | Bai, Jing Wang, Mengjie Kong, Dexin |
author_facet | Bai, Jing Wang, Mengjie Kong, Dexin |
author_sort | Bai, Jing |
collection | PubMed |
description | Sketch-based 3D model retrieval has become an important research topic in many applications, such as computer graphics and computer-aided design. Although sketches and 3D models have huge interdomain visual perception discrepancies, and sketches of the same object have remarkable intradomain visual perception diversity, the 3D models and sketches of the same class share common semantic content. Motivated by these findings, we propose a novel approach for sketch-based 3D model retrieval by constructing a deep common semantic space embedding using triplet network. First, a common data space is constructed by representing every 3D model as a group of views. Second, a common modality space is generated by translating views to sketches according to cross entropy evaluation. Third, a common semantic space embedding for two domains is learned based on a triplet network. Finally, based on the learned features of sketches and 3D models, four kinds of distance metrics between sketches and 3D models are designed, and sketch-based 3D model retrieval results are achieved. The experimental results using the Shape Retrieval Contest (SHREC) 2013 and SHREC 2014 datasets reveal the superiority of our proposed method over state-of-the-art methods. |
format | Online Article Text |
id | pubmed-7514853 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-75148532020-11-09 Deep Common Semantic Space Embedding for Sketch-Based 3D Model Retrieval Bai, Jing Wang, Mengjie Kong, Dexin Entropy (Basel) Article Sketch-based 3D model retrieval has become an important research topic in many applications, such as computer graphics and computer-aided design. Although sketches and 3D models have huge interdomain visual perception discrepancies, and sketches of the same object have remarkable intradomain visual perception diversity, the 3D models and sketches of the same class share common semantic content. Motivated by these findings, we propose a novel approach for sketch-based 3D model retrieval by constructing a deep common semantic space embedding using triplet network. First, a common data space is constructed by representing every 3D model as a group of views. Second, a common modality space is generated by translating views to sketches according to cross entropy evaluation. Third, a common semantic space embedding for two domains is learned based on a triplet network. Finally, based on the learned features of sketches and 3D models, four kinds of distance metrics between sketches and 3D models are designed, and sketch-based 3D model retrieval results are achieved. The experimental results using the Shape Retrieval Contest (SHREC) 2013 and SHREC 2014 datasets reveal the superiority of our proposed method over state-of-the-art methods. MDPI 2019-04-04 /pmc/articles/PMC7514853/ /pubmed/33267083 http://dx.doi.org/10.3390/e21040369 Text en © 2019 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Bai, Jing Wang, Mengjie Kong, Dexin Deep Common Semantic Space Embedding for Sketch-Based 3D Model Retrieval |
title | Deep Common Semantic Space Embedding for Sketch-Based 3D Model Retrieval |
title_full | Deep Common Semantic Space Embedding for Sketch-Based 3D Model Retrieval |
title_fullStr | Deep Common Semantic Space Embedding for Sketch-Based 3D Model Retrieval |
title_full_unstemmed | Deep Common Semantic Space Embedding for Sketch-Based 3D Model Retrieval |
title_short | Deep Common Semantic Space Embedding for Sketch-Based 3D Model Retrieval |
title_sort | deep common semantic space embedding for sketch-based 3d model retrieval |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7514853/ https://www.ncbi.nlm.nih.gov/pubmed/33267083 http://dx.doi.org/10.3390/e21040369 |
work_keys_str_mv | AT baijing deepcommonsemanticspaceembeddingforsketchbased3dmodelretrieval AT wangmengjie deepcommonsemanticspaceembeddingforsketchbased3dmodelretrieval AT kongdexin deepcommonsemanticspaceembeddingforsketchbased3dmodelretrieval |