Cargando…
Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model
Cross-modal hashing encodes heterogeneous multimedia data into compact binary code to achieve fast and flexible retrieval across different modalities. Due to its low storage cost and high retrieval efficiency, it has received widespread attention. Supervised deep hashing significantly improves searc...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Hindawi
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8310450/ https://www.ncbi.nlm.nih.gov/pubmed/34326867 http://dx.doi.org/10.1155/2021/5107034 |
_version_ | 1783728765386358784 |
---|---|
author | Li, Mingyong Li, Qiqi Tang, Lirong Peng, Shuang Ma, Yan Yang, Degang |
author_facet | Li, Mingyong Li, Qiqi Tang, Lirong Peng, Shuang Ma, Yan Yang, Degang |
author_sort | Li, Mingyong |
collection | PubMed |
description | Cross-modal hashing encodes heterogeneous multimedia data into compact binary code to achieve fast and flexible retrieval across different modalities. Due to its low storage cost and high retrieval efficiency, it has received widespread attention. Supervised deep hashing significantly improves search performance and usually yields more accurate results, but requires a lot of manual annotation of the data. In contrast, unsupervised deep hashing is difficult to achieve satisfactory performance due to the lack of reliable supervisory information. To solve this problem, inspired by knowledge distillation, we propose a novel unsupervised knowledge distillation cross-modal hashing method based on semantic alignment (SAKDH), which can reconstruct the similarity matrix using the hidden correlation information of the pretrained unsupervised teacher model, and the reconstructed similarity matrix can be used to guide the supervised student model. Specifically, firstly, the teacher model adopted an unsupervised semantic alignment hashing method, which can construct a modal fusion similarity matrix. Secondly, under the supervision of teacher model distillation information, the student model can generate more discriminative hash codes. Experimental results on two extensive benchmark datasets (MIRFLICKR-25K and NUS-WIDE) show that compared to several representative unsupervised cross-modal hashing methods, the mean average precision (MAP) of our proposed method has achieved a significant improvement. It fully reflects its effectiveness in large-scale cross-modal data retrieval. |
format | Online Article Text |
id | pubmed-8310450 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Hindawi |
record_format | MEDLINE/PubMed |
spelling | pubmed-83104502021-07-28 Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model Li, Mingyong Li, Qiqi Tang, Lirong Peng, Shuang Ma, Yan Yang, Degang Comput Intell Neurosci Research Article Cross-modal hashing encodes heterogeneous multimedia data into compact binary code to achieve fast and flexible retrieval across different modalities. Due to its low storage cost and high retrieval efficiency, it has received widespread attention. Supervised deep hashing significantly improves search performance and usually yields more accurate results, but requires a lot of manual annotation of the data. In contrast, unsupervised deep hashing is difficult to achieve satisfactory performance due to the lack of reliable supervisory information. To solve this problem, inspired by knowledge distillation, we propose a novel unsupervised knowledge distillation cross-modal hashing method based on semantic alignment (SAKDH), which can reconstruct the similarity matrix using the hidden correlation information of the pretrained unsupervised teacher model, and the reconstructed similarity matrix can be used to guide the supervised student model. Specifically, firstly, the teacher model adopted an unsupervised semantic alignment hashing method, which can construct a modal fusion similarity matrix. Secondly, under the supervision of teacher model distillation information, the student model can generate more discriminative hash codes. Experimental results on two extensive benchmark datasets (MIRFLICKR-25K and NUS-WIDE) show that compared to several representative unsupervised cross-modal hashing methods, the mean average precision (MAP) of our proposed method has achieved a significant improvement. It fully reflects its effectiveness in large-scale cross-modal data retrieval. Hindawi 2021-07-17 /pmc/articles/PMC8310450/ /pubmed/34326867 http://dx.doi.org/10.1155/2021/5107034 Text en Copyright © 2021 Mingyong Li et al. https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Research Article Li, Mingyong Li, Qiqi Tang, Lirong Peng, Shuang Ma, Yan Yang, Degang Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model |
title | Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model |
title_full | Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model |
title_fullStr | Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model |
title_full_unstemmed | Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model |
title_short | Deep Unsupervised Hashing for Large-Scale Cross-Modal Retrieval Using Knowledge Distillation Model |
title_sort | deep unsupervised hashing for large-scale cross-modal retrieval using knowledge distillation model |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8310450/ https://www.ncbi.nlm.nih.gov/pubmed/34326867 http://dx.doi.org/10.1155/2021/5107034 |
work_keys_str_mv | AT limingyong deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel AT liqiqi deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel AT tanglirong deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel AT pengshuang deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel AT mayan deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel AT yangdegang deepunsupervisedhashingforlargescalecrossmodalretrievalusingknowledgedistillationmodel |