Cargando…

An unsupervised learning approach to ultrasound strain elastography with spatio-temporal consistency

Quasi-static ultrasound elastography (USE) is an imaging modality that measures deformation (i.e. strain) of tissue in response to an applied mechanical force. In USE, the strain modulus is traditionally obtained by deriving the displacement field estimated between a pair of radio-frequency data. In...

Descripción completa

Detalles Bibliográficos
Autores principales: Delaunay, Rémi, Hu, Yipeng, Vercauteren, Tom
Formato: Online Artículo Texto
Lenguaje:English
Publicado: IOP Publishing 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8417818/
https://www.ncbi.nlm.nih.gov/pubmed/34298531
http://dx.doi.org/10.1088/1361-6560/ac176a
_version_ 1783748460054315008
author Delaunay, Rémi
Hu, Yipeng
Vercauteren, Tom
author_facet Delaunay, Rémi
Hu, Yipeng
Vercauteren, Tom
author_sort Delaunay, Rémi
collection PubMed
description Quasi-static ultrasound elastography (USE) is an imaging modality that measures deformation (i.e. strain) of tissue in response to an applied mechanical force. In USE, the strain modulus is traditionally obtained by deriving the displacement field estimated between a pair of radio-frequency data. In this work we propose a recurrent network architecture with convolutional long-short-term memory decoder blocks to improve displacement estimation and spatio-temporal continuity between time series ultrasound frames. The network is trained in an unsupervised way, by optimising a similarity metric between the reference and compressed image. Our training loss is also composed of a regularisation term that preserves displacement continuity by directly optimising the strain smoothness, and a temporal continuity term that enforces consistency between successive strain predictions. In addition, we propose an open-access in vivo database for quasi-static USE, which consists of radio-frequency data sequences captured on the arm of a human volunteer. Our results from numerical simulation and in vivo data suggest that our recurrent neural network can account for larger deformations, as compared with two other feed-forward neural networks. In all experiments, our recurrent network outperformed the state-of-the-art for both learning-based and optimisation-based methods, in terms of elastographic signal-to-noise ratio, strain consistency, and image similarity. Finally, our open-source code provides a 3D-slicer visualisation module that can be used to process ultrasound RF frames in real-time, at a rate of up to 20 frames per second, using a standard GPU.
format Online
Article
Text
id pubmed-8417818
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher IOP Publishing
record_format MEDLINE/PubMed
spelling pubmed-84178182021-09-07 An unsupervised learning approach to ultrasound strain elastography with spatio-temporal consistency Delaunay, Rémi Hu, Yipeng Vercauteren, Tom Phys Med Biol Paper Quasi-static ultrasound elastography (USE) is an imaging modality that measures deformation (i.e. strain) of tissue in response to an applied mechanical force. In USE, the strain modulus is traditionally obtained by deriving the displacement field estimated between a pair of radio-frequency data. In this work we propose a recurrent network architecture with convolutional long-short-term memory decoder blocks to improve displacement estimation and spatio-temporal continuity between time series ultrasound frames. The network is trained in an unsupervised way, by optimising a similarity metric between the reference and compressed image. Our training loss is also composed of a regularisation term that preserves displacement continuity by directly optimising the strain smoothness, and a temporal continuity term that enforces consistency between successive strain predictions. In addition, we propose an open-access in vivo database for quasi-static USE, which consists of radio-frequency data sequences captured on the arm of a human volunteer. Our results from numerical simulation and in vivo data suggest that our recurrent neural network can account for larger deformations, as compared with two other feed-forward neural networks. In all experiments, our recurrent network outperformed the state-of-the-art for both learning-based and optimisation-based methods, in terms of elastographic signal-to-noise ratio, strain consistency, and image similarity. Finally, our open-source code provides a 3D-slicer visualisation module that can be used to process ultrasound RF frames in real-time, at a rate of up to 20 frames per second, using a standard GPU. IOP Publishing 2021-09-07 2021-09-03 /pmc/articles/PMC8417818/ /pubmed/34298531 http://dx.doi.org/10.1088/1361-6560/ac176a Text en © 2021 Institute of Physics and Engineering in Medicine https://creativecommons.org/licenses/by/4.0/Original content from this work may be used under the terms of the Creative Commons Attribution 4.0 licence (https://creativecommons.org/licenses/by/4.0/) . Any further distribution of this work must maintain attribution to the author(s) and the title of the work, journal citation and DOI.
spellingShingle Paper
Delaunay, Rémi
Hu, Yipeng
Vercauteren, Tom
An unsupervised learning approach to ultrasound strain elastography with spatio-temporal consistency
title An unsupervised learning approach to ultrasound strain elastography with spatio-temporal consistency
title_full An unsupervised learning approach to ultrasound strain elastography with spatio-temporal consistency
title_fullStr An unsupervised learning approach to ultrasound strain elastography with spatio-temporal consistency
title_full_unstemmed An unsupervised learning approach to ultrasound strain elastography with spatio-temporal consistency
title_short An unsupervised learning approach to ultrasound strain elastography with spatio-temporal consistency
title_sort unsupervised learning approach to ultrasound strain elastography with spatio-temporal consistency
topic Paper
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8417818/
https://www.ncbi.nlm.nih.gov/pubmed/34298531
http://dx.doi.org/10.1088/1361-6560/ac176a
work_keys_str_mv AT delaunayremi anunsupervisedlearningapproachtoultrasoundstrainelastographywithspatiotemporalconsistency
AT huyipeng anunsupervisedlearningapproachtoultrasoundstrainelastographywithspatiotemporalconsistency
AT vercauterentom anunsupervisedlearningapproachtoultrasoundstrainelastographywithspatiotemporalconsistency
AT delaunayremi unsupervisedlearningapproachtoultrasoundstrainelastographywithspatiotemporalconsistency
AT huyipeng unsupervisedlearningapproachtoultrasoundstrainelastographywithspatiotemporalconsistency
AT vercauterentom unsupervisedlearningapproachtoultrasoundstrainelastographywithspatiotemporalconsistency