Cargando…

Entropy-Based Image Fusion with Joint Sparse Representation and Rolling Guidance Filter

Image fusion is a very practical technology that can be applied in many fields, such as medicine, remote sensing and surveillance. An image fusion method using multi-scale decomposition and joint sparse representation is introduced in this paper. First, joint sparse representation is applied to deco...

Descripción completa

Detalles Bibliográficos
Autores principales: Liu, Yudan, Yang, Xiaomin, Zhang, Rongzhu, Albertini, Marcelo Keese, Celik, Turgay, Jeon, Gwanggil
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516424/
https://www.ncbi.nlm.nih.gov/pubmed/33285893
http://dx.doi.org/10.3390/e22010118
_version_ 1783586997932130304
author Liu, Yudan
Yang, Xiaomin
Zhang, Rongzhu
Albertini, Marcelo Keese
Celik, Turgay
Jeon, Gwanggil
author_facet Liu, Yudan
Yang, Xiaomin
Zhang, Rongzhu
Albertini, Marcelo Keese
Celik, Turgay
Jeon, Gwanggil
author_sort Liu, Yudan
collection PubMed
description Image fusion is a very practical technology that can be applied in many fields, such as medicine, remote sensing and surveillance. An image fusion method using multi-scale decomposition and joint sparse representation is introduced in this paper. First, joint sparse representation is applied to decompose two source images into a common image and two innovation images. Second, two initial weight maps are generated by filtering the two source images separately. Final weight maps are obtained by joint bilateral filtering according to the initial weight maps. Then, the multi-scale decomposition of the innovation images is performed through the rolling guide filter. Finally, the final weight maps are used to generate the fused innovation image. The fused innovation image and the common image are combined to generate the ultimate fused image. The experimental results show that our method’s average metrics are: mutual information ([Formula: see text])—5.3377, feature mutual information ([Formula: see text])—0.5600, normalized weighted edge preservation value ([Formula: see text])—0.6978 and nonlinear correlation information entropy ([Formula: see text])—0.8226. Our method can achieve better performance compared to the state-of-the-art methods in visual perception and objective quantification.
format Online
Article
Text
id pubmed-7516424
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-75164242020-11-09 Entropy-Based Image Fusion with Joint Sparse Representation and Rolling Guidance Filter Liu, Yudan Yang, Xiaomin Zhang, Rongzhu Albertini, Marcelo Keese Celik, Turgay Jeon, Gwanggil Entropy (Basel) Article Image fusion is a very practical technology that can be applied in many fields, such as medicine, remote sensing and surveillance. An image fusion method using multi-scale decomposition and joint sparse representation is introduced in this paper. First, joint sparse representation is applied to decompose two source images into a common image and two innovation images. Second, two initial weight maps are generated by filtering the two source images separately. Final weight maps are obtained by joint bilateral filtering according to the initial weight maps. Then, the multi-scale decomposition of the innovation images is performed through the rolling guide filter. Finally, the final weight maps are used to generate the fused innovation image. The fused innovation image and the common image are combined to generate the ultimate fused image. The experimental results show that our method’s average metrics are: mutual information ([Formula: see text])—5.3377, feature mutual information ([Formula: see text])—0.5600, normalized weighted edge preservation value ([Formula: see text])—0.6978 and nonlinear correlation information entropy ([Formula: see text])—0.8226. Our method can achieve better performance compared to the state-of-the-art methods in visual perception and objective quantification. MDPI 2020-01-18 /pmc/articles/PMC7516424/ /pubmed/33285893 http://dx.doi.org/10.3390/e22010118 Text en © 2020 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Liu, Yudan
Yang, Xiaomin
Zhang, Rongzhu
Albertini, Marcelo Keese
Celik, Turgay
Jeon, Gwanggil
Entropy-Based Image Fusion with Joint Sparse Representation and Rolling Guidance Filter
title Entropy-Based Image Fusion with Joint Sparse Representation and Rolling Guidance Filter
title_full Entropy-Based Image Fusion with Joint Sparse Representation and Rolling Guidance Filter
title_fullStr Entropy-Based Image Fusion with Joint Sparse Representation and Rolling Guidance Filter
title_full_unstemmed Entropy-Based Image Fusion with Joint Sparse Representation and Rolling Guidance Filter
title_short Entropy-Based Image Fusion with Joint Sparse Representation and Rolling Guidance Filter
title_sort entropy-based image fusion with joint sparse representation and rolling guidance filter
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7516424/
https://www.ncbi.nlm.nih.gov/pubmed/33285893
http://dx.doi.org/10.3390/e22010118
work_keys_str_mv AT liuyudan entropybasedimagefusionwithjointsparserepresentationandrollingguidancefilter
AT yangxiaomin entropybasedimagefusionwithjointsparserepresentationandrollingguidancefilter
AT zhangrongzhu entropybasedimagefusionwithjointsparserepresentationandrollingguidancefilter
AT albertinimarcelokeese entropybasedimagefusionwithjointsparserepresentationandrollingguidancefilter
AT celikturgay entropybasedimagefusionwithjointsparserepresentationandrollingguidancefilter
AT jeongwanggil entropybasedimagefusionwithjointsparserepresentationandrollingguidancefilter