Cargando…

Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection

In harsh weather conditions, the infrared modality can supplement or even replace the visible modality. However, the lack of a large-scale dataset for infrared features hinders the generation of a robust pre-training model. Most existing infrared object-detection algorithms rely on pre-training mode...

Descripción completa

Detalles Bibliográficos
Autores principales: Wei, Chiheng, Bai, Lianfa, Chen, Xiaoyu, Han, Jing
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10453456/
https://www.ncbi.nlm.nih.gov/pubmed/37628196
http://dx.doi.org/10.3390/e25081166
_version_ 1785095940536795136
author Wei, Chiheng
Bai, Lianfa
Chen, Xiaoyu
Han, Jing
author_facet Wei, Chiheng
Bai, Lianfa
Chen, Xiaoyu
Han, Jing
author_sort Wei, Chiheng
collection PubMed
description In harsh weather conditions, the infrared modality can supplement or even replace the visible modality. However, the lack of a large-scale dataset for infrared features hinders the generation of a robust pre-training model. Most existing infrared object-detection algorithms rely on pre-training models from the visible modality, which can accelerate network convergence but also limit performance due to modality differences. In order to provide more reliable feature representation for cross-modality object detection and enhance its performance, this paper investigates the impact of various task-relevant features on cross-modality object detection and proposes a knowledge transfer algorithm based on classification and localization decoupling analysis. A task-decoupled pre-training method is introduced to adjust the attributes of various tasks learned by the pre-training model. For the training phase, a task-relevant hyperparameter evolution method is proposed to increase the network’s adaptability to attribute changes in pre-training weights. Our proposed method improves the accuracy of multiple modalities in multiple datasets, with experimental results on the FLIR ADAS dataset reaching a state-of-the-art level and surpassing most multi-spectral object-detection methods.
format Online
Article
Text
id pubmed-10453456
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-104534562023-08-26 Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection Wei, Chiheng Bai, Lianfa Chen, Xiaoyu Han, Jing Entropy (Basel) Article In harsh weather conditions, the infrared modality can supplement or even replace the visible modality. However, the lack of a large-scale dataset for infrared features hinders the generation of a robust pre-training model. Most existing infrared object-detection algorithms rely on pre-training models from the visible modality, which can accelerate network convergence but also limit performance due to modality differences. In order to provide more reliable feature representation for cross-modality object detection and enhance its performance, this paper investigates the impact of various task-relevant features on cross-modality object detection and proposes a knowledge transfer algorithm based on classification and localization decoupling analysis. A task-decoupled pre-training method is introduced to adjust the attributes of various tasks learned by the pre-training model. For the training phase, a task-relevant hyperparameter evolution method is proposed to increase the network’s adaptability to attribute changes in pre-training weights. Our proposed method improves the accuracy of multiple modalities in multiple datasets, with experimental results on the FLIR ADAS dataset reaching a state-of-the-art level and surpassing most multi-spectral object-detection methods. MDPI 2023-08-04 /pmc/articles/PMC10453456/ /pubmed/37628196 http://dx.doi.org/10.3390/e25081166 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Wei, Chiheng
Bai, Lianfa
Chen, Xiaoyu
Han, Jing
Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection
title Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection
title_full Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection
title_fullStr Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection
title_full_unstemmed Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection
title_short Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection
title_sort task-decoupled knowledge transfer for cross-modality object detection
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10453456/
https://www.ncbi.nlm.nih.gov/pubmed/37628196
http://dx.doi.org/10.3390/e25081166
work_keys_str_mv AT weichiheng taskdecoupledknowledgetransferforcrossmodalityobjectdetection
AT bailianfa taskdecoupledknowledgetransferforcrossmodalityobjectdetection
AT chenxiaoyu taskdecoupledknowledgetransferforcrossmodalityobjectdetection
AT hanjing taskdecoupledknowledgetransferforcrossmodalityobjectdetection