Cargando…

Deep Learning to Obtain Simultaneous Image and Segmentation Outputs From a Single Input of Raw Ultrasound Channel Data

Single plane wave transmissions are promising for automated imaging tasks requiring high ultrasound frame rates over an extended field of view. However, a single plane wave insonification typically produces suboptimal image quality. To address this limitation, we are exploring the use of deep neural...

Descripción completa

Detalles Bibliográficos
Autores principales: Nair, Arun Asokan, Washington, Kendra N., Tran, Trac D., Reiter, Austin, Bell, Muyinatu A. Lediju
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7990652/
https://www.ncbi.nlm.nih.gov/pubmed/32396084
http://dx.doi.org/10.1109/TUFFC.2020.2993779
_version_ 1783669109413642240
author Nair, Arun Asokan
Washington, Kendra N.
Tran, Trac D.
Reiter, Austin
Bell, Muyinatu A. Lediju
author_facet Nair, Arun Asokan
Washington, Kendra N.
Tran, Trac D.
Reiter, Austin
Bell, Muyinatu A. Lediju
author_sort Nair, Arun Asokan
collection PubMed
description Single plane wave transmissions are promising for automated imaging tasks requiring high ultrasound frame rates over an extended field of view. However, a single plane wave insonification typically produces suboptimal image quality. To address this limitation, we are exploring the use of deep neural networks (DNNs) as an alternative to delay-and-sum (DAS) beamforming. The objectives of this work are to obtain information directly from raw channel data and to simultaneously generate both a segmentation map for automated ultrasound tasks and a corresponding ultrasound B-mode image for interpretable supervision of the automation. We focus on visualizing and segmenting anechoic targets surrounded by tissue and ignoring or deemphasizing less important surrounding structures. DNNs trained with Field II simulations were tested with simulated, experimental phantom, and in vivo data sets that were not included during training. With unfocused input channel data (i.e., prior to the application of receive time delays), simulated, experimental phantom, and in vivo test data sets achieved mean ± standard deviation Dice similarity coefficients of 0.92 ± 0.13, 0.92 ± 0.03, and 0.77 ± 0.07, respectively, and generalized contrast-to-noise ratios (gCNRs) of 0.95 ± 0.08, 0.93 ± 0.08, and 0.75 ± 0.14, respectively. With subaperture beamformed channel data and a modification to the input layer of the DNN architecture to accept these data, the fidelity of image reconstruction increased (e.g., mean gCNR of multiple acquisitions of two in vivo breast cysts ranged 0.89–0.96), but DNN display frame rates were reduced from 395 to 287 Hz. Overall, the DNNs successfully translated feature representations learned from simulated data to phantom and in vivo data, which is promising for this novel approach to simultaneous ultrasound image formation and segmentation.
format Online
Article
Text
id pubmed-7990652
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-79906522021-03-25 Deep Learning to Obtain Simultaneous Image and Segmentation Outputs From a Single Input of Raw Ultrasound Channel Data Nair, Arun Asokan Washington, Kendra N. Tran, Trac D. Reiter, Austin Bell, Muyinatu A. Lediju IEEE Trans Ultrason Ferroelectr Freq Control Article Single plane wave transmissions are promising for automated imaging tasks requiring high ultrasound frame rates over an extended field of view. However, a single plane wave insonification typically produces suboptimal image quality. To address this limitation, we are exploring the use of deep neural networks (DNNs) as an alternative to delay-and-sum (DAS) beamforming. The objectives of this work are to obtain information directly from raw channel data and to simultaneously generate both a segmentation map for automated ultrasound tasks and a corresponding ultrasound B-mode image for interpretable supervision of the automation. We focus on visualizing and segmenting anechoic targets surrounded by tissue and ignoring or deemphasizing less important surrounding structures. DNNs trained with Field II simulations were tested with simulated, experimental phantom, and in vivo data sets that were not included during training. With unfocused input channel data (i.e., prior to the application of receive time delays), simulated, experimental phantom, and in vivo test data sets achieved mean ± standard deviation Dice similarity coefficients of 0.92 ± 0.13, 0.92 ± 0.03, and 0.77 ± 0.07, respectively, and generalized contrast-to-noise ratios (gCNRs) of 0.95 ± 0.08, 0.93 ± 0.08, and 0.75 ± 0.14, respectively. With subaperture beamformed channel data and a modification to the input layer of the DNN architecture to accept these data, the fidelity of image reconstruction increased (e.g., mean gCNR of multiple acquisitions of two in vivo breast cysts ranged 0.89–0.96), but DNN display frame rates were reduced from 395 to 287 Hz. Overall, the DNNs successfully translated feature representations learned from simulated data to phantom and in vivo data, which is promising for this novel approach to simultaneous ultrasound image formation and segmentation. 2020-11-24 2020-12 /pmc/articles/PMC7990652/ /pubmed/32396084 http://dx.doi.org/10.1109/TUFFC.2020.2993779 Text en This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
spellingShingle Article
Nair, Arun Asokan
Washington, Kendra N.
Tran, Trac D.
Reiter, Austin
Bell, Muyinatu A. Lediju
Deep Learning to Obtain Simultaneous Image and Segmentation Outputs From a Single Input of Raw Ultrasound Channel Data
title Deep Learning to Obtain Simultaneous Image and Segmentation Outputs From a Single Input of Raw Ultrasound Channel Data
title_full Deep Learning to Obtain Simultaneous Image and Segmentation Outputs From a Single Input of Raw Ultrasound Channel Data
title_fullStr Deep Learning to Obtain Simultaneous Image and Segmentation Outputs From a Single Input of Raw Ultrasound Channel Data
title_full_unstemmed Deep Learning to Obtain Simultaneous Image and Segmentation Outputs From a Single Input of Raw Ultrasound Channel Data
title_short Deep Learning to Obtain Simultaneous Image and Segmentation Outputs From a Single Input of Raw Ultrasound Channel Data
title_sort deep learning to obtain simultaneous image and segmentation outputs from a single input of raw ultrasound channel data
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7990652/
https://www.ncbi.nlm.nih.gov/pubmed/32396084
http://dx.doi.org/10.1109/TUFFC.2020.2993779
work_keys_str_mv AT nairarunasokan deeplearningtoobtainsimultaneousimageandsegmentationoutputsfromasingleinputofrawultrasoundchanneldata
AT washingtonkendran deeplearningtoobtainsimultaneousimageandsegmentationoutputsfromasingleinputofrawultrasoundchanneldata
AT trantracd deeplearningtoobtainsimultaneousimageandsegmentationoutputsfromasingleinputofrawultrasoundchanneldata
AT reiteraustin deeplearningtoobtainsimultaneousimageandsegmentationoutputsfromasingleinputofrawultrasoundchanneldata
AT bellmuyinatualediju deeplearningtoobtainsimultaneousimageandsegmentationoutputsfromasingleinputofrawultrasoundchanneldata