Cargando…

Infer Thermal Information from Visual Information: A Cross Imaging Modality Edge Learning (CIMEL) Framework

The measurement accuracy and reliability of thermography is largely limited by a relatively low spatial-resolution of infrared (IR) cameras in comparison to digital cameras. Using a high-end IR camera to achieve high spatial-resolution can be costly or sometimes infeasible due to the high sample rat...

Descripción completa

Detalles Bibliográficos
Autores principales: Wang, Shuozhi, Mei, Jianqiang, Yang, Lichao, Zhao, Yifan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8617953/
https://www.ncbi.nlm.nih.gov/pubmed/34833548
http://dx.doi.org/10.3390/s21227471
_version_ 1784604630835003392
author Wang, Shuozhi
Mei, Jianqiang
Yang, Lichao
Zhao, Yifan
author_facet Wang, Shuozhi
Mei, Jianqiang
Yang, Lichao
Zhao, Yifan
author_sort Wang, Shuozhi
collection PubMed
description The measurement accuracy and reliability of thermography is largely limited by a relatively low spatial-resolution of infrared (IR) cameras in comparison to digital cameras. Using a high-end IR camera to achieve high spatial-resolution can be costly or sometimes infeasible due to the high sample rate required. Therefore, there is a strong demand to improve the quality of IR images, particularly on edges, without upgrading the hardware in the context of surveillance and industrial inspection systems. This paper proposes a novel Conditional Generative Adversarial Networks (CGAN)-based framework to enhance IR edges by learning high-frequency features from corresponding visual images. A dual-discriminator, focusing on edge and content/background, is introduced to guide the cross imaging modality learning procedure of the U-Net generator in high and low frequencies respectively. Results demonstrate that the proposed framework can effectively enhance barely visible edges in IR images without introducing artefacts, meanwhile the content information is well preserved. Different from most similar studies, this method only requires IR images for testing, which will increase the applicability of some scenarios where only one imaging modality is available, such as active thermography.
format Online
Article
Text
id pubmed-8617953
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-86179532021-11-27 Infer Thermal Information from Visual Information: A Cross Imaging Modality Edge Learning (CIMEL) Framework Wang, Shuozhi Mei, Jianqiang Yang, Lichao Zhao, Yifan Sensors (Basel) Article The measurement accuracy and reliability of thermography is largely limited by a relatively low spatial-resolution of infrared (IR) cameras in comparison to digital cameras. Using a high-end IR camera to achieve high spatial-resolution can be costly or sometimes infeasible due to the high sample rate required. Therefore, there is a strong demand to improve the quality of IR images, particularly on edges, without upgrading the hardware in the context of surveillance and industrial inspection systems. This paper proposes a novel Conditional Generative Adversarial Networks (CGAN)-based framework to enhance IR edges by learning high-frequency features from corresponding visual images. A dual-discriminator, focusing on edge and content/background, is introduced to guide the cross imaging modality learning procedure of the U-Net generator in high and low frequencies respectively. Results demonstrate that the proposed framework can effectively enhance barely visible edges in IR images without introducing artefacts, meanwhile the content information is well preserved. Different from most similar studies, this method only requires IR images for testing, which will increase the applicability of some scenarios where only one imaging modality is available, such as active thermography. MDPI 2021-11-10 /pmc/articles/PMC8617953/ /pubmed/34833548 http://dx.doi.org/10.3390/s21227471 Text en © 2021 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Wang, Shuozhi
Mei, Jianqiang
Yang, Lichao
Zhao, Yifan
Infer Thermal Information from Visual Information: A Cross Imaging Modality Edge Learning (CIMEL) Framework
title Infer Thermal Information from Visual Information: A Cross Imaging Modality Edge Learning (CIMEL) Framework
title_full Infer Thermal Information from Visual Information: A Cross Imaging Modality Edge Learning (CIMEL) Framework
title_fullStr Infer Thermal Information from Visual Information: A Cross Imaging Modality Edge Learning (CIMEL) Framework
title_full_unstemmed Infer Thermal Information from Visual Information: A Cross Imaging Modality Edge Learning (CIMEL) Framework
title_short Infer Thermal Information from Visual Information: A Cross Imaging Modality Edge Learning (CIMEL) Framework
title_sort infer thermal information from visual information: a cross imaging modality edge learning (cimel) framework
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8617953/
https://www.ncbi.nlm.nih.gov/pubmed/34833548
http://dx.doi.org/10.3390/s21227471
work_keys_str_mv AT wangshuozhi inferthermalinformationfromvisualinformationacrossimagingmodalityedgelearningcimelframework
AT meijianqiang inferthermalinformationfromvisualinformationacrossimagingmodalityedgelearningcimelframework
AT yanglichao inferthermalinformationfromvisualinformationacrossimagingmodalityedgelearningcimelframework
AT zhaoyifan inferthermalinformationfromvisualinformationacrossimagingmodalityedgelearningcimelframework