Cargando…
Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning
In this work, we exploit multi-task learning to jointly predict the two decision-making processes of gaze movement and probe manipulation that an experienced sonographer would perform in routine obstetric scanning. A multimodal guidance framework, Multimodal-GuideNet, is proposed to detect the causa...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7615231/ https://www.ncbi.nlm.nih.gov/pubmed/37863638 http://dx.doi.org/10.1016/j.media.2023.102981 |
_version_ | 1785145787637825536 |
---|---|
author | Men, Qianhui Teng, Clare Drukker, Lior Papageorghiou, Aris T. Noble, J. Alison |
author_facet | Men, Qianhui Teng, Clare Drukker, Lior Papageorghiou, Aris T. Noble, J. Alison |
author_sort | Men, Qianhui |
collection | PubMed |
description | In this work, we exploit multi-task learning to jointly predict the two decision-making processes of gaze movement and probe manipulation that an experienced sonographer would perform in routine obstetric scanning. A multimodal guidance framework, Multimodal-GuideNet, is proposed to detect the causal relationship between a real-world ultrasound video signal, synchronized gaze, and probe motion. The association between the multi-modality inputs is learned and shared through a modality-aware spatial graph that leverages useful cross-modal dependencies. By estimating the probability distribution of probe and gaze movements in real scans, the predicted guidance signals also allow inter- and intra-sonographer variations and avoid a fixed scanning path. We validate the new multi-modality approach on three types of obstetric scanning examinations, and the result consistently outperforms single-task learning under various guidance policies. To simulate sonographer’s attention on multi-structure images, we also explore multi-step estimation in gaze guidance, and its visual results show that the prediction allows multiple gaze centers that are substantially aligned with underlying anatomical structures. |
format | Online Article Text |
id | pubmed-7615231 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
record_format | MEDLINE/PubMed |
spelling | pubmed-76152312023-12-01 Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning Men, Qianhui Teng, Clare Drukker, Lior Papageorghiou, Aris T. Noble, J. Alison Med Image Anal Article In this work, we exploit multi-task learning to jointly predict the two decision-making processes of gaze movement and probe manipulation that an experienced sonographer would perform in routine obstetric scanning. A multimodal guidance framework, Multimodal-GuideNet, is proposed to detect the causal relationship between a real-world ultrasound video signal, synchronized gaze, and probe motion. The association between the multi-modality inputs is learned and shared through a modality-aware spatial graph that leverages useful cross-modal dependencies. By estimating the probability distribution of probe and gaze movements in real scans, the predicted guidance signals also allow inter- and intra-sonographer variations and avoid a fixed scanning path. We validate the new multi-modality approach on three types of obstetric scanning examinations, and the result consistently outperforms single-task learning under various guidance policies. To simulate sonographer’s attention on multi-structure images, we also explore multi-step estimation in gaze guidance, and its visual results show that the prediction allows multiple gaze centers that are substantially aligned with underlying anatomical structures. 2023-12 /pmc/articles/PMC7615231/ /pubmed/37863638 http://dx.doi.org/10.1016/j.media.2023.102981 Text en https://creativecommons.org/licenses/by/4.0/This work is licensed under a CC BY 4.0 International license (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Men, Qianhui Teng, Clare Drukker, Lior Papageorghiou, Aris T. Noble, J. Alison Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning |
title | Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning |
title_full | Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning |
title_fullStr | Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning |
title_full_unstemmed | Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning |
title_short | Gaze-Probe Joint Guidance with Multi-task Learning in Obstetric Ultrasound Scanning |
title_sort | gaze-probe joint guidance with multi-task learning in obstetric ultrasound scanning |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7615231/ https://www.ncbi.nlm.nih.gov/pubmed/37863638 http://dx.doi.org/10.1016/j.media.2023.102981 |
work_keys_str_mv | AT menqianhui gazeprobejointguidancewithmultitasklearninginobstetricultrasoundscanning AT tengclare gazeprobejointguidancewithmultitasklearninginobstetricultrasoundscanning AT drukkerlior gazeprobejointguidancewithmultitasklearninginobstetricultrasoundscanning AT papageorghiouarist gazeprobejointguidancewithmultitasklearninginobstetricultrasoundscanning AT noblejalison gazeprobejointguidancewithmultitasklearninginobstetricultrasoundscanning |