Cargando…

Pay attention to doctor–patient dialogues: Multi-modal knowledge graph attention image-text embedding for COVID-19 diagnosis

The sudden increase in coronavirus disease 2019 (COVID-19) cases puts high pressure on healthcare services worldwide. At this stage, fast, accurate, and early clinical assessment of the disease severity is vital. In general, there are two issues to overcome: (1) Current deep learning-based works suf...

Descripción completa

Detalles Bibliográficos
Autores principales: Zheng, Wenbo, Yan, Lan, Gou, Chao, Zhang, Zhi-Cheng, Jason Zhang, Jun, Hu, Ming, Wang, Fei-Yue
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier B.V. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8168340/
https://www.ncbi.nlm.nih.gov/pubmed/34093095
http://dx.doi.org/10.1016/j.inffus.2021.05.015
_version_ 1783701868600360960
author Zheng, Wenbo
Yan, Lan
Gou, Chao
Zhang, Zhi-Cheng
Jason Zhang, Jun
Hu, Ming
Wang, Fei-Yue
author_facet Zheng, Wenbo
Yan, Lan
Gou, Chao
Zhang, Zhi-Cheng
Jason Zhang, Jun
Hu, Ming
Wang, Fei-Yue
author_sort Zheng, Wenbo
collection PubMed
description The sudden increase in coronavirus disease 2019 (COVID-19) cases puts high pressure on healthcare services worldwide. At this stage, fast, accurate, and early clinical assessment of the disease severity is vital. In general, there are two issues to overcome: (1) Current deep learning-based works suffer from multimodal data adequacy issues; (2) In this scenario, multimodal (e.g., text, image) information should be taken into account together to make accurate inferences. To address these challenges, we propose a multi-modal knowledge graph attention embedding for COVID-19 diagnosis. Our method not only learns the relational embedding from nodes in a constituted knowledge graph but also has access to medical knowledge, aiming at improving the performance of the classifier through the mechanism of medical knowledge attention. The experimental results show that our approach significantly improves classification performance compared to other state-of-the-art techniques and possesses robustness for each modality from multi-modal data. Moreover, we construct a new COVID-19 multi-modal dataset based on text mining, consisting of 1393 doctor–patient dialogues and their 3706 images (347 X-ray [Formula: see text] 2598 CT [Formula: see text] 761 ultrasound) about COVID-19 patients and 607 non-COVID-19 patient dialogues and their 10754 images (9658 X-ray [Formula: see text] 494 CT [Formula: see text] 761 ultrasound), and the fine-grained labels of all. We hope this work can provide insights to the researchers working in this area to shift the attention from only medical images to the doctor–patient dialogue and its corresponding medical images.
format Online
Article
Text
id pubmed-8168340
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Elsevier B.V.
record_format MEDLINE/PubMed
spelling pubmed-81683402021-06-01 Pay attention to doctor–patient dialogues: Multi-modal knowledge graph attention image-text embedding for COVID-19 diagnosis Zheng, Wenbo Yan, Lan Gou, Chao Zhang, Zhi-Cheng Jason Zhang, Jun Hu, Ming Wang, Fei-Yue Inf Fusion Article The sudden increase in coronavirus disease 2019 (COVID-19) cases puts high pressure on healthcare services worldwide. At this stage, fast, accurate, and early clinical assessment of the disease severity is vital. In general, there are two issues to overcome: (1) Current deep learning-based works suffer from multimodal data adequacy issues; (2) In this scenario, multimodal (e.g., text, image) information should be taken into account together to make accurate inferences. To address these challenges, we propose a multi-modal knowledge graph attention embedding for COVID-19 diagnosis. Our method not only learns the relational embedding from nodes in a constituted knowledge graph but also has access to medical knowledge, aiming at improving the performance of the classifier through the mechanism of medical knowledge attention. The experimental results show that our approach significantly improves classification performance compared to other state-of-the-art techniques and possesses robustness for each modality from multi-modal data. Moreover, we construct a new COVID-19 multi-modal dataset based on text mining, consisting of 1393 doctor–patient dialogues and their 3706 images (347 X-ray [Formula: see text] 2598 CT [Formula: see text] 761 ultrasound) about COVID-19 patients and 607 non-COVID-19 patient dialogues and their 10754 images (9658 X-ray [Formula: see text] 494 CT [Formula: see text] 761 ultrasound), and the fine-grained labels of all. We hope this work can provide insights to the researchers working in this area to shift the attention from only medical images to the doctor–patient dialogue and its corresponding medical images. Elsevier B.V. 2021-11 2021-06-01 /pmc/articles/PMC8168340/ /pubmed/34093095 http://dx.doi.org/10.1016/j.inffus.2021.05.015 Text en © 2021 Elsevier B.V. All rights reserved. Since January 2020 Elsevier has created a COVID-19 resource centre with free information in English and Mandarin on the novel coronavirus COVID-19. The COVID-19 resource centre is hosted on Elsevier Connect, the company's public news and information website. Elsevier hereby grants permission to make all its COVID-19-related research that is available on the COVID-19 resource centre - including this research content - immediately available in PubMed Central and other publicly funded repositories, such as the WHO COVID database with rights for unrestricted research re-use and analyses in any form or by any means with acknowledgement of the original source. These permissions are granted for free by Elsevier for as long as the COVID-19 resource centre remains active.
spellingShingle Article
Zheng, Wenbo
Yan, Lan
Gou, Chao
Zhang, Zhi-Cheng
Jason Zhang, Jun
Hu, Ming
Wang, Fei-Yue
Pay attention to doctor–patient dialogues: Multi-modal knowledge graph attention image-text embedding for COVID-19 diagnosis
title Pay attention to doctor–patient dialogues: Multi-modal knowledge graph attention image-text embedding for COVID-19 diagnosis
title_full Pay attention to doctor–patient dialogues: Multi-modal knowledge graph attention image-text embedding for COVID-19 diagnosis
title_fullStr Pay attention to doctor–patient dialogues: Multi-modal knowledge graph attention image-text embedding for COVID-19 diagnosis
title_full_unstemmed Pay attention to doctor–patient dialogues: Multi-modal knowledge graph attention image-text embedding for COVID-19 diagnosis
title_short Pay attention to doctor–patient dialogues: Multi-modal knowledge graph attention image-text embedding for COVID-19 diagnosis
title_sort pay attention to doctor–patient dialogues: multi-modal knowledge graph attention image-text embedding for covid-19 diagnosis
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8168340/
https://www.ncbi.nlm.nih.gov/pubmed/34093095
http://dx.doi.org/10.1016/j.inffus.2021.05.015
work_keys_str_mv AT zhengwenbo payattentiontodoctorpatientdialoguesmultimodalknowledgegraphattentionimagetextembeddingforcovid19diagnosis
AT yanlan payattentiontodoctorpatientdialoguesmultimodalknowledgegraphattentionimagetextembeddingforcovid19diagnosis
AT gouchao payattentiontodoctorpatientdialoguesmultimodalknowledgegraphattentionimagetextembeddingforcovid19diagnosis
AT zhangzhicheng payattentiontodoctorpatientdialoguesmultimodalknowledgegraphattentionimagetextembeddingforcovid19diagnosis
AT jasonzhangjun payattentiontodoctorpatientdialoguesmultimodalknowledgegraphattentionimagetextembeddingforcovid19diagnosis
AT huming payattentiontodoctorpatientdialoguesmultimodalknowledgegraphattentionimagetextembeddingforcovid19diagnosis
AT wangfeiyue payattentiontodoctorpatientdialoguesmultimodalknowledgegraphattentionimagetextembeddingforcovid19diagnosis