Cargando…
A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech
OBJECTIVE: Dynamic magnetic resonance (MR) imaging enables visualisation of articulators during speech. There is growing interest in quantifying articulator motion in two-dimensional MR images of the vocal tract, to better understand speech production and potentially inform patient management decisi...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Elsevier
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9746295/ https://www.ncbi.nlm.nih.gov/pubmed/36743699 http://dx.doi.org/10.1016/j.bspc.2022.104290 |
_version_ | 1784849330242322432 |
---|---|
author | Ruthven, Matthieu Miquel, Marc E. King, Andrew P. |
author_facet | Ruthven, Matthieu Miquel, Marc E. King, Andrew P. |
author_sort | Ruthven, Matthieu |
collection | PubMed |
description | OBJECTIVE: Dynamic magnetic resonance (MR) imaging enables visualisation of articulators during speech. There is growing interest in quantifying articulator motion in two-dimensional MR images of the vocal tract, to better understand speech production and potentially inform patient management decisions. Image registration is an established way to achieve this quantification. Recently, segmentation-informed deformable registration frameworks have been developed and have achieved state-of-the-art accuracy. This work aims to adapt such a framework and optimise it for estimating displacement fields between dynamic two-dimensional MR images of the vocal tract during speech. METHODS: A deep-learning-based registration framework was developed and compared with current state-of-the-art registration methods and frameworks (two traditional methods and three deep-learning-based frameworks, two of which are segmentation informed). The accuracy of the methods and frameworks was evaluated using the Dice coefficient (DSC), average surface distance (ASD) and a metric based on velopharyngeal closure. The metric evaluated if the fields captured a clinically relevant and quantifiable aspect of articulator motion. RESULTS: The segmentation-informed frameworks achieved higher DSCs and lower ASDs and captured more velopharyngeal closures than the traditional methods and the framework that was not segmentation informed. All segmentation-informed frameworks achieved similar DSCs and ASDs. However, the proposed framework captured the most velopharyngeal closures. CONCLUSIONS: A framework was successfully developed and found to more accurately estimate articulator motion than five current state-of-the-art methods and frameworks. SIGNIFICANCE: The first deep-learning-based framework specifically for registering dynamic two-dimensional MR images of the vocal tract during speech has been developed and evaluated. |
format | Online Article Text |
id | pubmed-9746295 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Elsevier |
record_format | MEDLINE/PubMed |
spelling | pubmed-97462952023-02-01 A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech Ruthven, Matthieu Miquel, Marc E. King, Andrew P. Biomed Signal Process Control Article OBJECTIVE: Dynamic magnetic resonance (MR) imaging enables visualisation of articulators during speech. There is growing interest in quantifying articulator motion in two-dimensional MR images of the vocal tract, to better understand speech production and potentially inform patient management decisions. Image registration is an established way to achieve this quantification. Recently, segmentation-informed deformable registration frameworks have been developed and have achieved state-of-the-art accuracy. This work aims to adapt such a framework and optimise it for estimating displacement fields between dynamic two-dimensional MR images of the vocal tract during speech. METHODS: A deep-learning-based registration framework was developed and compared with current state-of-the-art registration methods and frameworks (two traditional methods and three deep-learning-based frameworks, two of which are segmentation informed). The accuracy of the methods and frameworks was evaluated using the Dice coefficient (DSC), average surface distance (ASD) and a metric based on velopharyngeal closure. The metric evaluated if the fields captured a clinically relevant and quantifiable aspect of articulator motion. RESULTS: The segmentation-informed frameworks achieved higher DSCs and lower ASDs and captured more velopharyngeal closures than the traditional methods and the framework that was not segmentation informed. All segmentation-informed frameworks achieved similar DSCs and ASDs. However, the proposed framework captured the most velopharyngeal closures. CONCLUSIONS: A framework was successfully developed and found to more accurately estimate articulator motion than five current state-of-the-art methods and frameworks. SIGNIFICANCE: The first deep-learning-based framework specifically for registering dynamic two-dimensional MR images of the vocal tract during speech has been developed and evaluated. Elsevier 2023-02 /pmc/articles/PMC9746295/ /pubmed/36743699 http://dx.doi.org/10.1016/j.bspc.2022.104290 Text en © 2022 The Author(s) https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/). |
spellingShingle | Article Ruthven, Matthieu Miquel, Marc E. King, Andrew P. A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech |
title | A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech |
title_full | A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech |
title_fullStr | A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech |
title_full_unstemmed | A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech |
title_short | A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech |
title_sort | segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9746295/ https://www.ncbi.nlm.nih.gov/pubmed/36743699 http://dx.doi.org/10.1016/j.bspc.2022.104290 |
work_keys_str_mv | AT ruthvenmatthieu asegmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech AT miquelmarce asegmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech AT kingandrewp asegmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech AT ruthvenmatthieu segmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech AT miquelmarce segmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech AT kingandrewp segmentationinformeddeeplearningframeworktoregisterdynamictwodimensionalmagneticresonanceimagesofthevocaltractduringspeech |