Cargando…

Deep learning autofluorescence-harmonic microscopy

Laser scanning microscopy has inherent tradeoffs between imaging speed, field of view (FOV), and spatial resolution due to the limitations of sophisticated mechanical and optical setups, and deep learning networks have emerged to overcome these limitations without changing the system. Here, we demon...

Descripción completa

Detalles Bibliográficos
Autores principales: Shen, Binglin, Liu, Shaowen, Li, Yanping, Pan, Ying, Lu, Yuan, Hu, Rui, Qu, Junle, Liu, Liwei
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8964717/
https://www.ncbi.nlm.nih.gov/pubmed/35351853
http://dx.doi.org/10.1038/s41377-022-00768-x
_version_ 1784678279624523776
author Shen, Binglin
Liu, Shaowen
Li, Yanping
Pan, Ying
Lu, Yuan
Hu, Rui
Qu, Junle
Liu, Liwei
author_facet Shen, Binglin
Liu, Shaowen
Li, Yanping
Pan, Ying
Lu, Yuan
Hu, Rui
Qu, Junle
Liu, Liwei
author_sort Shen, Binglin
collection PubMed
description Laser scanning microscopy has inherent tradeoffs between imaging speed, field of view (FOV), and spatial resolution due to the limitations of sophisticated mechanical and optical setups, and deep learning networks have emerged to overcome these limitations without changing the system. Here, we demonstrate deep learning autofluorescence-harmonic microscopy (DLAM) based on self-alignment attention-guided residual-in-residual dense generative adversarial networks to close the gap between speed, FOV, and quality. Using the framework, we demonstrate label-free large-field multimodal imaging of clinicopathological tissues with enhanced spatial resolution and running time advantages. Statistical quality assessments show that the attention-guided residual dense connections minimize the persistent noise, distortions, and scanning fringes that degrade the autofluorescence-harmonic images and avoid reconstruction artifacts in the output images. With the advantages of high contrast, high fidelity, and high speed in image reconstruction, DLAM can act as a powerful tool for the noninvasive evaluation of diseases, neural activity, and embryogenesis.
format Online
Article
Text
id pubmed-8964717
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-89647172022-04-12 Deep learning autofluorescence-harmonic microscopy Shen, Binglin Liu, Shaowen Li, Yanping Pan, Ying Lu, Yuan Hu, Rui Qu, Junle Liu, Liwei Light Sci Appl Article Laser scanning microscopy has inherent tradeoffs between imaging speed, field of view (FOV), and spatial resolution due to the limitations of sophisticated mechanical and optical setups, and deep learning networks have emerged to overcome these limitations without changing the system. Here, we demonstrate deep learning autofluorescence-harmonic microscopy (DLAM) based on self-alignment attention-guided residual-in-residual dense generative adversarial networks to close the gap between speed, FOV, and quality. Using the framework, we demonstrate label-free large-field multimodal imaging of clinicopathological tissues with enhanced spatial resolution and running time advantages. Statistical quality assessments show that the attention-guided residual dense connections minimize the persistent noise, distortions, and scanning fringes that degrade the autofluorescence-harmonic images and avoid reconstruction artifacts in the output images. With the advantages of high contrast, high fidelity, and high speed in image reconstruction, DLAM can act as a powerful tool for the noninvasive evaluation of diseases, neural activity, and embryogenesis. Nature Publishing Group UK 2022-03-29 /pmc/articles/PMC8964717/ /pubmed/35351853 http://dx.doi.org/10.1038/s41377-022-00768-x Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Shen, Binglin
Liu, Shaowen
Li, Yanping
Pan, Ying
Lu, Yuan
Hu, Rui
Qu, Junle
Liu, Liwei
Deep learning autofluorescence-harmonic microscopy
title Deep learning autofluorescence-harmonic microscopy
title_full Deep learning autofluorescence-harmonic microscopy
title_fullStr Deep learning autofluorescence-harmonic microscopy
title_full_unstemmed Deep learning autofluorescence-harmonic microscopy
title_short Deep learning autofluorescence-harmonic microscopy
title_sort deep learning autofluorescence-harmonic microscopy
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8964717/
https://www.ncbi.nlm.nih.gov/pubmed/35351853
http://dx.doi.org/10.1038/s41377-022-00768-x
work_keys_str_mv AT shenbinglin deeplearningautofluorescenceharmonicmicroscopy
AT liushaowen deeplearningautofluorescenceharmonicmicroscopy
AT liyanping deeplearningautofluorescenceharmonicmicroscopy
AT panying deeplearningautofluorescenceharmonicmicroscopy
AT luyuan deeplearningautofluorescenceharmonicmicroscopy
AT hurui deeplearningautofluorescenceharmonicmicroscopy
AT qujunle deeplearningautofluorescenceharmonicmicroscopy
AT liuliwei deeplearningautofluorescenceharmonicmicroscopy