Cargando…

A Novel Infrared and Visible Image Fusion Approach Based on Adversarial Neural Network

The presence of fake pictures affects the reliability of visible face images under specific circumstances. This paper presents a novel adversarial neural network designed named as the FTSGAN for infrared and visible image fusion and we utilize FTSGAN model to fuse the face image features of infrared...

Descripción completa

Detalles Bibliográficos
Autores principales: Chen, Xianglong, Wang, Haipeng, Liang, Yaohui, Meng, Ying, Wang, Shifeng
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8749719/
https://www.ncbi.nlm.nih.gov/pubmed/35009852
http://dx.doi.org/10.3390/s22010304
_version_ 1784631296607125504
author Chen, Xianglong
Wang, Haipeng
Liang, Yaohui
Meng, Ying
Wang, Shifeng
author_facet Chen, Xianglong
Wang, Haipeng
Liang, Yaohui
Meng, Ying
Wang, Shifeng
author_sort Chen, Xianglong
collection PubMed
description The presence of fake pictures affects the reliability of visible face images under specific circumstances. This paper presents a novel adversarial neural network designed named as the FTSGAN for infrared and visible image fusion and we utilize FTSGAN model to fuse the face image features of infrared and visible image to improve the effect of face recognition. In FTSGAN model design, the Frobenius norm (F), total variation norm (TV), and structural similarity index measure (SSIM) are employed. The F and TV are used to limit the gray level and the gradient of the image, while the SSIM is used to limit the image structure. The FTSGAN fuses infrared and visible face images that contains bio-information for heterogeneous face recognition tasks. Experiments based on the FTSGAN using hundreds of face images demonstrate its excellent performance. The principal component analysis (PCA) and linear discrimination analysis (LDA) are involved in face recognition. The face recognition performance after fusion improved by 1.9% compared to that before fusion, and the final face recognition rate was 94.4%. This proposed method has better quality, faster rate, and is more robust than the methods that only use visible images for face recognition.
format Online
Article
Text
id pubmed-8749719
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-87497192022-01-12 A Novel Infrared and Visible Image Fusion Approach Based on Adversarial Neural Network Chen, Xianglong Wang, Haipeng Liang, Yaohui Meng, Ying Wang, Shifeng Sensors (Basel) Article The presence of fake pictures affects the reliability of visible face images under specific circumstances. This paper presents a novel adversarial neural network designed named as the FTSGAN for infrared and visible image fusion and we utilize FTSGAN model to fuse the face image features of infrared and visible image to improve the effect of face recognition. In FTSGAN model design, the Frobenius norm (F), total variation norm (TV), and structural similarity index measure (SSIM) are employed. The F and TV are used to limit the gray level and the gradient of the image, while the SSIM is used to limit the image structure. The FTSGAN fuses infrared and visible face images that contains bio-information for heterogeneous face recognition tasks. Experiments based on the FTSGAN using hundreds of face images demonstrate its excellent performance. The principal component analysis (PCA) and linear discrimination analysis (LDA) are involved in face recognition. The face recognition performance after fusion improved by 1.9% compared to that before fusion, and the final face recognition rate was 94.4%. This proposed method has better quality, faster rate, and is more robust than the methods that only use visible images for face recognition. MDPI 2021-12-31 /pmc/articles/PMC8749719/ /pubmed/35009852 http://dx.doi.org/10.3390/s22010304 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Chen, Xianglong
Wang, Haipeng
Liang, Yaohui
Meng, Ying
Wang, Shifeng
A Novel Infrared and Visible Image Fusion Approach Based on Adversarial Neural Network
title A Novel Infrared and Visible Image Fusion Approach Based on Adversarial Neural Network
title_full A Novel Infrared and Visible Image Fusion Approach Based on Adversarial Neural Network
title_fullStr A Novel Infrared and Visible Image Fusion Approach Based on Adversarial Neural Network
title_full_unstemmed A Novel Infrared and Visible Image Fusion Approach Based on Adversarial Neural Network
title_short A Novel Infrared and Visible Image Fusion Approach Based on Adversarial Neural Network
title_sort novel infrared and visible image fusion approach based on adversarial neural network
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8749719/
https://www.ncbi.nlm.nih.gov/pubmed/35009852
http://dx.doi.org/10.3390/s22010304
work_keys_str_mv AT chenxianglong anovelinfraredandvisibleimagefusionapproachbasedonadversarialneuralnetwork
AT wanghaipeng anovelinfraredandvisibleimagefusionapproachbasedonadversarialneuralnetwork
AT liangyaohui anovelinfraredandvisibleimagefusionapproachbasedonadversarialneuralnetwork
AT mengying anovelinfraredandvisibleimagefusionapproachbasedonadversarialneuralnetwork
AT wangshifeng anovelinfraredandvisibleimagefusionapproachbasedonadversarialneuralnetwork
AT chenxianglong novelinfraredandvisibleimagefusionapproachbasedonadversarialneuralnetwork
AT wanghaipeng novelinfraredandvisibleimagefusionapproachbasedonadversarialneuralnetwork
AT liangyaohui novelinfraredandvisibleimagefusionapproachbasedonadversarialneuralnetwork
AT mengying novelinfraredandvisibleimagefusionapproachbasedonadversarialneuralnetwork
AT wangshifeng novelinfraredandvisibleimagefusionapproachbasedonadversarialneuralnetwork