Cargando…

Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning

In this work, we exploit multi-task learning to jointly predict the two decision-making processes of gaze movement and probe manipulation that an experienced sonographer would perform in routine obstetric scanning. A multimodal guidance framework, Multimodal-GuideNet, is proposed to detect the causa...

Descripción completa

Detalles Bibliográficos
Autores principales: Men, Qianhui, Teng, Clare, Drukker, Lior, Papageorghiou, Aris T., Noble, J. Alison
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7615231/
https://www.ncbi.nlm.nih.gov/pubmed/37863638
http://dx.doi.org/10.1016/j.media.2023.102981
_version_ 1785145787637825536
author Men, Qianhui
Teng, Clare
Drukker, Lior
Papageorghiou, Aris T.
Noble, J. Alison
author_facet Men, Qianhui
Teng, Clare
Drukker, Lior
Papageorghiou, Aris T.
Noble, J. Alison
author_sort Men, Qianhui
collection PubMed
description In this work, we exploit multi-task learning to jointly predict the two decision-making processes of gaze movement and probe manipulation that an experienced sonographer would perform in routine obstetric scanning. A multimodal guidance framework, Multimodal-GuideNet, is proposed to detect the causal relationship between a real-world ultrasound video signal, synchronized gaze, and probe motion. The association between the multi-modality inputs is learned and shared through a modality-aware spatial graph that leverages useful cross-modal dependencies. By estimating the probability distribution of probe and gaze movements in real scans, the predicted guidance signals also allow inter- and intra-sonographer variations and avoid a fixed scanning path. We validate the new multi-modality approach on three types of obstetric scanning examinations, and the result consistently outperforms single-task learning under various guidance policies. To simulate sonographer’s attention on multi-structure images, we also explore multi-step estimation in gaze guidance, and its visual results show that the prediction allows multiple gaze centers that are substantially aligned with underlying anatomical structures.
format Online
Article
Text
id pubmed-7615231
institution National Center for Biotechnology Information
language English
publishDate 2023
record_format MEDLINE/PubMed
spelling pubmed-76152312023-12-01 Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning Men, Qianhui Teng, Clare Drukker, Lior Papageorghiou, Aris T. Noble, J. Alison Med Image Anal Article In this work, we exploit multi-task learning to jointly predict the two decision-making processes of gaze movement and probe manipulation that an experienced sonographer would perform in routine obstetric scanning. A multimodal guidance framework, Multimodal-GuideNet, is proposed to detect the causal relationship between a real-world ultrasound video signal, synchronized gaze, and probe motion. The association between the multi-modality inputs is learned and shared through a modality-aware spatial graph that leverages useful cross-modal dependencies. By estimating the probability distribution of probe and gaze movements in real scans, the predicted guidance signals also allow inter- and intra-sonographer variations and avoid a fixed scanning path. We validate the new multi-modality approach on three types of obstetric scanning examinations, and the result consistently outperforms single-task learning under various guidance policies. To simulate sonographer’s attention on multi-structure images, we also explore multi-step estimation in gaze guidance, and its visual results show that the prediction allows multiple gaze centers that are substantially aligned with underlying anatomical structures. 2023-12 /pmc/articles/PMC7615231/ /pubmed/37863638 http://dx.doi.org/10.1016/j.media.2023.102981 Text en https://creativecommons.org/licenses/by/4.0/This work is licensed under a CC BY 4.0 International license (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Men, Qianhui
Teng, Clare
Drukker, Lior
Papageorghiou, Aris T.
Noble, J. Alison
Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning
title Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning
title_full Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning
title_fullStr Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning
title_full_unstemmed Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning
title_short Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning
title_sort gaze-probe joint guidance with multi-task learning in obstetric ultrasound scanning
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7615231/
https://www.ncbi.nlm.nih.gov/pubmed/37863638
http://dx.doi.org/10.1016/j.media.2023.102981
work_keys_str_mv AT menqianhui gazeprobejointguidancewithmultitasklearninginobstetricultrasoundscanning
AT tengclare gazeprobejointguidancewithmultitasklearninginobstetricultrasoundscanning
AT drukkerlior gazeprobejointguidancewithmultitasklearninginobstetricultrasoundscanning
AT papageorghiouarist gazeprobejointguidancewithmultitasklearninginobstetricultrasoundscanning
AT noblejalison gazeprobejointguidancewithmultitasklearninginobstetricultrasoundscanning