Cargando…

A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech

OBJECTIVE: Dynamic magnetic resonance (MR) imaging enables visualisation of articulators during speech. There is growing interest in quantifying articulator motion in two-dimensional MR images of the vocal tract, to better understand speech production and potentially inform patient management decisi...

Descripción completa

Detalles Bibliográficos
Autores principales: Ruthven, Matthieu, Miquel, Marc E., King, Andrew P.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9746295/
https://www.ncbi.nlm.nih.gov/pubmed/36743699
http://dx.doi.org/10.1016/j.bspc.2022.104290
_version_ 1784849330242322432
author Ruthven, Matthieu
Miquel, Marc E.
King, Andrew P.
author_facet Ruthven, Matthieu
Miquel, Marc E.
King, Andrew P.
author_sort Ruthven, Matthieu
collection PubMed
description OBJECTIVE: Dynamic magnetic resonance (MR) imaging enables visualisation of articulators during speech. There is growing interest in quantifying articulator motion in two-dimensional MR images of the vocal tract, to better understand speech production and potentially inform patient management decisions. Image registration is an established way to achieve this quantification. Recently, segmentation-informed deformable registration frameworks have been developed and have achieved state-of-the-art accuracy. This work aims to adapt such a framework and optimise it for estimating displacement fields between dynamic two-dimensional MR images of the vocal tract during speech. METHODS: A deep-learning-based registration framework was developed and compared with current state-of-the-art registration methods and frameworks (two traditional methods and three deep-learning-based frameworks, two of which are segmentation informed). The accuracy of the methods and frameworks was evaluated using the Dice coefficient (DSC), average surface distance (ASD) and a metric based on velopharyngeal closure. The metric evaluated if the fields captured a clinically relevant and quantifiable aspect of articulator motion. RESULTS: The segmentation-informed frameworks achieved higher DSCs and lower ASDs and captured more velopharyngeal closures than the traditional methods and the framework that was not segmentation informed. All segmentation-informed frameworks achieved similar DSCs and ASDs. However, the proposed framework captured the most velopharyngeal closures. CONCLUSIONS: A framework was successfully developed and found to more accurately estimate articulator motion than five current state-of-the-art methods and frameworks. SIGNIFICANCE: The first deep-learning-based framework specifically for registering dynamic two-dimensional MR images of the vocal tract during speech has been developed and evaluated.
format Online
Article
Text
id pubmed-9746295
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Elsevier
record_format MEDLINE/PubMed
spelling pubmed-97462952023-02-01 A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech Ruthven, Matthieu Miquel, Marc E. King, Andrew P. Biomed Signal Process Control Article OBJECTIVE: Dynamic magnetic resonance (MR) imaging enables visualisation of articulators during speech. There is growing interest in quantifying articulator motion in two-dimensional MR images of the vocal tract, to better understand speech production and potentially inform patient management decisions. Image registration is an established way to achieve this quantification. Recently, segmentation-informed deformable registration frameworks have been developed and have achieved state-of-the-art accuracy. This work aims to adapt such a framework and optimise it for estimating displacement fields between dynamic two-dimensional MR images of the vocal tract during speech. METHODS: A deep-learning-based registration framework was developed and compared with current state-of-the-art registration methods and frameworks (two traditional methods and three deep-learning-based frameworks, two of which are segmentation informed). The accuracy of the methods and frameworks was evaluated using the Dice coefficient (DSC), average surface distance (ASD) and a metric based on velopharyngeal closure. The metric evaluated if the fields captured a clinically relevant and quantifiable aspect of articulator motion. RESULTS: The segmentation-informed frameworks achieved higher DSCs and lower ASDs and captured more velopharyngeal closures than the traditional methods and the framework that was not segmentation informed. All segmentation-informed frameworks achieved similar DSCs and ASDs. However, the proposed framework captured the most velopharyngeal closures. CONCLUSIONS: A framework was successfully developed and found to more accurately estimate articulator motion than five current state-of-the-art methods and frameworks. SIGNIFICANCE: The first deep-learning-based framework specifically for registering dynamic two-dimensional MR images of the vocal tract during speech has been developed and evaluated. Elsevier 2023-02 /pmc/articles/PMC9746295/ /pubmed/36743699 http://dx.doi.org/10.1016/j.bspc.2022.104290 Text en © 2022 The Author(s) https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Ruthven, Matthieu
Miquel, Marc E.
King, Andrew P.
A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech
title A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech
title_full A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech
title_fullStr A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech
title_full_unstemmed A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech
title_short A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech
title_sort segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9746295/
https://www.ncbi.nlm.nih.gov/pubmed/36743699
http://dx.doi.org/10.1016/j.bspc.2022.104290
work_keys_str_mv AT ruthvenmatthieu asegmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech
AT miquelmarce asegmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech
AT kingandrewp asegmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech
AT ruthvenmatthieu segmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech
AT miquelmarce segmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech
AT kingandrewp segmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech