Cargando…
Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection
In harsh weather conditions, the infrared modality can supplement or even replace the visible modality. However, the lack of a large-scale dataset for infrared features hinders the generation of a robust pre-training model. Most existing infrared object-detection algorithms rely on pre-training mode...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10453456/ https://www.ncbi.nlm.nih.gov/pubmed/37628196 http://dx.doi.org/10.3390/e25081166 |
_version_ | 1785095940536795136 |
---|---|
author | Wei, Chiheng Bai, Lianfa Chen, Xiaoyu Han, Jing |
author_facet | Wei, Chiheng Bai, Lianfa Chen, Xiaoyu Han, Jing |
author_sort | Wei, Chiheng |
collection | PubMed |
description | In harsh weather conditions, the infrared modality can supplement or even replace the visible modality. However, the lack of a large-scale dataset for infrared features hinders the generation of a robust pre-training model. Most existing infrared object-detection algorithms rely on pre-training models from the visible modality, which can accelerate network convergence but also limit performance due to modality differences. In order to provide more reliable feature representation for cross-modality object detection and enhance its performance, this paper investigates the impact of various task-relevant features on cross-modality object detection and proposes a knowledge transfer algorithm based on classification and localization decoupling analysis. A task-decoupled pre-training method is introduced to adjust the attributes of various tasks learned by the pre-training model. For the training phase, a task-relevant hyperparameter evolution method is proposed to increase the network’s adaptability to attribute changes in pre-training weights. Our proposed method improves the accuracy of multiple modalities in multiple datasets, with experimental results on the FLIR ADAS dataset reaching a state-of-the-art level and surpassing most multi-spectral object-detection methods. |
format | Online Article Text |
id | pubmed-10453456 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-104534562023-08-26 Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection Wei, Chiheng Bai, Lianfa Chen, Xiaoyu Han, Jing Entropy (Basel) Article In harsh weather conditions, the infrared modality can supplement or even replace the visible modality. However, the lack of a large-scale dataset for infrared features hinders the generation of a robust pre-training model. Most existing infrared object-detection algorithms rely on pre-training models from the visible modality, which can accelerate network convergence but also limit performance due to modality differences. In order to provide more reliable feature representation for cross-modality object detection and enhance its performance, this paper investigates the impact of various task-relevant features on cross-modality object detection and proposes a knowledge transfer algorithm based on classification and localization decoupling analysis. A task-decoupled pre-training method is introduced to adjust the attributes of various tasks learned by the pre-training model. For the training phase, a task-relevant hyperparameter evolution method is proposed to increase the network’s adaptability to attribute changes in pre-training weights. Our proposed method improves the accuracy of multiple modalities in multiple datasets, with experimental results on the FLIR ADAS dataset reaching a state-of-the-art level and surpassing most multi-spectral object-detection methods. MDPI 2023-08-04 /pmc/articles/PMC10453456/ /pubmed/37628196 http://dx.doi.org/10.3390/e25081166 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Wei, Chiheng Bai, Lianfa Chen, Xiaoyu Han, Jing Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection |
title | Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection |
title_full | Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection |
title_fullStr | Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection |
title_full_unstemmed | Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection |
title_short | Task-Decoupled Knowledge Transfer for Cross-Modality Object Detection |
title_sort | task-decoupled knowledge transfer for cross-modality object detection |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10453456/ https://www.ncbi.nlm.nih.gov/pubmed/37628196 http://dx.doi.org/10.3390/e25081166 |
work_keys_str_mv | AT weichiheng taskdecoupledknowledgetransferforcrossmodalityobjectdetection AT bailianfa taskdecoupledknowledgetransferforcrossmodalityobjectdetection AT chenxiaoyu taskdecoupledknowledgetransferforcrossmodalityobjectdetection AT hanjing taskdecoupledknowledgetransferforcrossmodalityobjectdetection |