Cargando…

DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution

Infrared and visible image fusion aims to generate a single fused image that not only contains rich texture details and salient objects, but also facilitates downstream tasks. However, existing works mainly focus on learning different modality-specific or shared features, and ignore the importance o...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhu, Huayi, Wu, Heshan, Wang, Xiaolong, He, Dongmei, Liu, Zhenbing, Pan, Xipeng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10458385/
https://www.ncbi.nlm.nih.gov/pubmed/37631742
http://dx.doi.org/10.3390/s23167205
_version_ 1785097152967475200
author Zhu, Huayi
Wu, Heshan
Wang, Xiaolong
He, Dongmei
Liu, Zhenbing
Pan, Xipeng
author_facet Zhu, Huayi
Wu, Heshan
Wang, Xiaolong
He, Dongmei
Liu, Zhenbing
Pan, Xipeng
author_sort Zhu, Huayi
collection PubMed
description Infrared and visible image fusion aims to generate a single fused image that not only contains rich texture details and salient objects, but also facilitates downstream tasks. However, existing works mainly focus on learning different modality-specific or shared features, and ignore the importance of modeling cross-modality features. To address these challenges, we propose Dual-branch Progressive learning for infrared and visible image fusion with a complementary self-Attention and Convolution (DPACFuse) network. On the one hand, we propose Cross-Modality Feature Extraction (CMEF) to enhance information interaction and the extraction of common features across modalities. In addition, we introduce a high-frequency gradient convolution operation to extract fine-grained information and suppress high-frequency information loss. On the other hand, to alleviate the CNN issues of insufficient global information extraction and computation overheads of self-attention, we introduce the ACmix, which can fully extract local and global information in the source image with a smaller computational overhead than pure convolution or pure self-attention. Extensive experiments demonstrated that the fused images generated by DPACFuse not only contain rich texture information, but can also effectively highlight salient objects. Additionally, our method achieved approximately 3% improvement over the state-of-the-art methods in MI, Qabf, SF, and AG evaluation indicators. More importantly, our fused images enhanced object detection and semantic segmentation by approximately 10%, compared to using infrared and visible images separately.
format Online
Article
Text
id pubmed-10458385
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-104583852023-08-27 DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution Zhu, Huayi Wu, Heshan Wang, Xiaolong He, Dongmei Liu, Zhenbing Pan, Xipeng Sensors (Basel) Article Infrared and visible image fusion aims to generate a single fused image that not only contains rich texture details and salient objects, but also facilitates downstream tasks. However, existing works mainly focus on learning different modality-specific or shared features, and ignore the importance of modeling cross-modality features. To address these challenges, we propose Dual-branch Progressive learning for infrared and visible image fusion with a complementary self-Attention and Convolution (DPACFuse) network. On the one hand, we propose Cross-Modality Feature Extraction (CMEF) to enhance information interaction and the extraction of common features across modalities. In addition, we introduce a high-frequency gradient convolution operation to extract fine-grained information and suppress high-frequency information loss. On the other hand, to alleviate the CNN issues of insufficient global information extraction and computation overheads of self-attention, we introduce the ACmix, which can fully extract local and global information in the source image with a smaller computational overhead than pure convolution or pure self-attention. Extensive experiments demonstrated that the fused images generated by DPACFuse not only contain rich texture information, but can also effectively highlight salient objects. Additionally, our method achieved approximately 3% improvement over the state-of-the-art methods in MI, Qabf, SF, and AG evaluation indicators. More importantly, our fused images enhanced object detection and semantic segmentation by approximately 10%, compared to using infrared and visible images separately. MDPI 2023-08-16 /pmc/articles/PMC10458385/ /pubmed/37631742 http://dx.doi.org/10.3390/s23167205 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhu, Huayi
Wu, Heshan
Wang, Xiaolong
He, Dongmei
Liu, Zhenbing
Pan, Xipeng
DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution
title DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution
title_full DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution
title_fullStr DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution
title_full_unstemmed DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution
title_short DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution
title_sort dpacfuse: dual-branch progressive learning for infrared and visible image fusion with complementary self-attention and convolution
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10458385/
https://www.ncbi.nlm.nih.gov/pubmed/37631742
http://dx.doi.org/10.3390/s23167205
work_keys_str_mv AT zhuhuayi dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution
AT wuheshan dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution
AT wangxiaolong dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution
AT hedongmei dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution
AT liuzhenbing dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution
AT panxipeng dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution