Cargando…
DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution
Infrared and visible image fusion aims to generate a single fused image that not only contains rich texture details and salient objects, but also facilitates downstream tasks. However, existing works mainly focus on learning different modality-specific or shared features, and ignore the importance o...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
MDPI
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10458385/ https://www.ncbi.nlm.nih.gov/pubmed/37631742 http://dx.doi.org/10.3390/s23167205 |
_version_ | 1785097152967475200 |
---|---|
author | Zhu, Huayi Wu, Heshan Wang, Xiaolong He, Dongmei Liu, Zhenbing Pan, Xipeng |
author_facet | Zhu, Huayi Wu, Heshan Wang, Xiaolong He, Dongmei Liu, Zhenbing Pan, Xipeng |
author_sort | Zhu, Huayi |
collection | PubMed |
description | Infrared and visible image fusion aims to generate a single fused image that not only contains rich texture details and salient objects, but also facilitates downstream tasks. However, existing works mainly focus on learning different modality-specific or shared features, and ignore the importance of modeling cross-modality features. To address these challenges, we propose Dual-branch Progressive learning for infrared and visible image fusion with a complementary self-Attention and Convolution (DPACFuse) network. On the one hand, we propose Cross-Modality Feature Extraction (CMEF) to enhance information interaction and the extraction of common features across modalities. In addition, we introduce a high-frequency gradient convolution operation to extract fine-grained information and suppress high-frequency information loss. On the other hand, to alleviate the CNN issues of insufficient global information extraction and computation overheads of self-attention, we introduce the ACmix, which can fully extract local and global information in the source image with a smaller computational overhead than pure convolution or pure self-attention. Extensive experiments demonstrated that the fused images generated by DPACFuse not only contain rich texture information, but can also effectively highlight salient objects. Additionally, our method achieved approximately 3% improvement over the state-of-the-art methods in MI, Qabf, SF, and AG evaluation indicators. More importantly, our fused images enhanced object detection and semantic segmentation by approximately 10%, compared to using infrared and visible images separately. |
format | Online Article Text |
id | pubmed-10458385 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | MDPI |
record_format | MEDLINE/PubMed |
spelling | pubmed-104583852023-08-27 DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution Zhu, Huayi Wu, Heshan Wang, Xiaolong He, Dongmei Liu, Zhenbing Pan, Xipeng Sensors (Basel) Article Infrared and visible image fusion aims to generate a single fused image that not only contains rich texture details and salient objects, but also facilitates downstream tasks. However, existing works mainly focus on learning different modality-specific or shared features, and ignore the importance of modeling cross-modality features. To address these challenges, we propose Dual-branch Progressive learning for infrared and visible image fusion with a complementary self-Attention and Convolution (DPACFuse) network. On the one hand, we propose Cross-Modality Feature Extraction (CMEF) to enhance information interaction and the extraction of common features across modalities. In addition, we introduce a high-frequency gradient convolution operation to extract fine-grained information and suppress high-frequency information loss. On the other hand, to alleviate the CNN issues of insufficient global information extraction and computation overheads of self-attention, we introduce the ACmix, which can fully extract local and global information in the source image with a smaller computational overhead than pure convolution or pure self-attention. Extensive experiments demonstrated that the fused images generated by DPACFuse not only contain rich texture information, but can also effectively highlight salient objects. Additionally, our method achieved approximately 3% improvement over the state-of-the-art methods in MI, Qabf, SF, and AG evaluation indicators. More importantly, our fused images enhanced object detection and semantic segmentation by approximately 10%, compared to using infrared and visible images separately. MDPI 2023-08-16 /pmc/articles/PMC10458385/ /pubmed/37631742 http://dx.doi.org/10.3390/s23167205 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Zhu, Huayi Wu, Heshan Wang, Xiaolong He, Dongmei Liu, Zhenbing Pan, Xipeng DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution |
title | DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution |
title_full | DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution |
title_fullStr | DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution |
title_full_unstemmed | DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution |
title_short | DPACFuse: Dual-Branch Progressive Learning for Infrared and Visible Image Fusion with Complementary Self-Attention and Convolution |
title_sort | dpacfuse: dual-branch progressive learning for infrared and visible image fusion with complementary self-attention and convolution |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10458385/ https://www.ncbi.nlm.nih.gov/pubmed/37631742 http://dx.doi.org/10.3390/s23167205 |
work_keys_str_mv | AT zhuhuayi dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution AT wuheshan dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution AT wangxiaolong dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution AT hedongmei dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution AT liuzhenbing dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution AT panxipeng dpacfusedualbranchprogressivelearningforinfraredandvisibleimagefusionwithcomplementaryselfattentionandconvolution |