Cargando…

Toward accurate quantitative photoacoustic imaging: learning vascular blood oxygen saturation in three dimensions

Significance: Two-dimensional (2-D) fully convolutional neural networks have been shown capable of producing maps of [Formula: see text] from 2-D simulated images of simple tissue models. However, their potential to produce accurate estimates in vivo is uncertain as they are limited by the 2-D natur...

Descripción completa

Detalles Bibliográficos
Autores principales: Bench, Ciaran, Hauptmann, Andreas, Cox, Ben
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Society of Photo-Optical Instrumentation Engineers 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7443711/
https://www.ncbi.nlm.nih.gov/pubmed/32840068
http://dx.doi.org/10.1117/1.JBO.25.8.085003
Descripción
Sumario:Significance: Two-dimensional (2-D) fully convolutional neural networks have been shown capable of producing maps of [Formula: see text] from 2-D simulated images of simple tissue models. However, their potential to produce accurate estimates in vivo is uncertain as they are limited by the 2-D nature of the training data when the problem is inherently three-dimensional (3-D), and they have not been tested with realistic images. Aim: To demonstrate the capability of deep neural networks to process whole 3-D images and output 3-D maps of vascular [Formula: see text] from realistic tissue models/images. Approach: Two separate fully convolutional neural networks were trained to produce 3-D maps of vascular blood oxygen saturation and vessel positions from multiwavelength simulated images of tissue models. Results: The mean of the absolute difference between the true mean vessel [Formula: see text] and the network output for 40 examples was 4.4% and the standard deviation was 4.5%. Conclusions: 3-D fully convolutional networks were shown capable of producing accurate [Formula: see text] maps using the full extent of spatial information contained within 3-D images generated under conditions mimicking real imaging scenarios. We demonstrate that networks can cope with some of the confounding effects present in real images such as limited-view artifacts and have the potential to produce accurate estimates in vivo.