Cargando…

Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model

Cross-modal hashing encodes heterogeneous multimedia data into compact binary code to achieve fast and flexible retrieval across different modalities. Due to its low storage cost and high retrieval efficiency, it has received widespread attention. Supervised deep hashing significantly improves searc...

Descripción completa

Detalles Bibliográficos
Autores principales: Li, Mingyong, Li, Qiqi, Tang, Lirong, Peng, Shuang, Ma, Yan, Yang, Degang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Hindawi 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8310450/
https://www.ncbi.nlm.nih.gov/pubmed/34326867
http://dx.doi.org/10.1155/2021/5107034
_version_ 1783728765386358784
author Li, Mingyong
Li, Qiqi
Tang, Lirong
Peng, Shuang
Ma, Yan
Yang, Degang
author_facet Li, Mingyong
Li, Qiqi
Tang, Lirong
Peng, Shuang
Ma, Yan
Yang, Degang
author_sort Li, Mingyong
collection PubMed
description Cross-modal hashing encodes heterogeneous multimedia data into compact binary code to achieve fast and flexible retrieval across different modalities. Due to its low storage cost and high retrieval efficiency, it has received widespread attention. Supervised deep hashing significantly improves search performance and usually yields more accurate results, but requires a lot of manual annotation of the data. In contrast, unsupervised deep hashing is difficult to achieve satisfactory performance due to the lack of reliable supervisory information. To solve this problem, inspired by knowledge distillation, we propose a novel unsupervised knowledge distillation cross-modal hashing method based on semantic alignment (SAKDH), which can reconstruct the similarity matrix using the hidden correlation information of the pretrained unsupervised teacher model, and the reconstructed similarity matrix can be used to guide the supervised student model. Specifically, firstly, the teacher model adopted an unsupervised semantic alignment hashing method, which can construct a modal fusion similarity matrix. Secondly, under the supervision of teacher model distillation information, the student model can generate more discriminative hash codes. Experimental results on two extensive benchmark datasets (MIRFLICKR-25K and NUS-WIDE) show that compared to several representative unsupervised cross-modal hashing methods, the mean average precision (MAP) of our proposed method has achieved a significant improvement. It fully reflects its effectiveness in large-scale cross-modal data retrieval.
format Online
Article
Text
id pubmed-8310450
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Hindawi
record_format MEDLINE/PubMed
spelling pubmed-83104502021-07-28 Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model Li, Mingyong Li, Qiqi Tang, Lirong Peng, Shuang Ma, Yan Yang, Degang Comput Intell Neurosci Research Article Cross-modal hashing encodes heterogeneous multimedia data into compact binary code to achieve fast and flexible retrieval across different modalities. Due to its low storage cost and high retrieval efficiency, it has received widespread attention. Supervised deep hashing significantly improves search performance and usually yields more accurate results, but requires a lot of manual annotation of the data. In contrast, unsupervised deep hashing is difficult to achieve satisfactory performance due to the lack of reliable supervisory information. To solve this problem, inspired by knowledge distillation, we propose a novel unsupervised knowledge distillation cross-modal hashing method based on semantic alignment (SAKDH), which can reconstruct the similarity matrix using the hidden correlation information of the pretrained unsupervised teacher model, and the reconstructed similarity matrix can be used to guide the supervised student model. Specifically, firstly, the teacher model adopted an unsupervised semantic alignment hashing method, which can construct a modal fusion similarity matrix. Secondly, under the supervision of teacher model distillation information, the student model can generate more discriminative hash codes. Experimental results on two extensive benchmark datasets (MIRFLICKR-25K and NUS-WIDE) show that compared to several representative unsupervised cross-modal hashing methods, the mean average precision (MAP) of our proposed method has achieved a significant improvement. It fully reflects its effectiveness in large-scale cross-modal data retrieval. Hindawi 2021-07-17 /pmc/articles/PMC8310450/ /pubmed/34326867 http://dx.doi.org/10.1155/2021/5107034 Text en Copyright © 2021 Mingyong Li et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
spellingShingle Research Article
Li, Mingyong
Li, Qiqi
Tang, Lirong
Peng, Shuang
Ma, Yan
Yang, Degang
Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model
title Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model
title_full Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model
title_fullStr Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model
title_full_unstemmed Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model
title_short Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model
title_sort deep unsupervised hashing for large-scale cross-modal retrieval using knowledge distillation model
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8310450/
https://www.ncbi.nlm.nih.gov/pubmed/34326867
http://dx.doi.org/10.1155/2021/5107034
work_keys_str_mv AT limingyong deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel
AT liqiqi deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel
AT tanglirong deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel
AT pengshuang deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel
AT mayan deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel
AT yangdegang deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel