Cargando…

Diversity Learning Based on Multi-Latent Space for Medical Image Visual Question Generation

Auxiliary clinical diagnosis has been researched to solve unevenly and insufficiently distributed clinical resources. However, auxiliary diagnosis is still dominated by human physicians, and how to make intelligent systems more involved in the diagnosis process is gradually becoming a concern. An in...

Descripción completa

Detalles Bibliográficos
Autores principales: Zhu, He, Togo, Ren, Ogawa, Takahiro, Haseyama, Miki
Formato: Online Artículo Texto
Lenguaje:English
Publicado: MDPI 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9919063/
https://www.ncbi.nlm.nih.gov/pubmed/36772095
http://dx.doi.org/10.3390/s23031057
_version_ 1784886731275763712
author Zhu, He
Togo, Ren
Ogawa, Takahiro
Haseyama, Miki
author_facet Zhu, He
Togo, Ren
Ogawa, Takahiro
Haseyama, Miki
author_sort Zhu, He
collection PubMed
description Auxiliary clinical diagnosis has been researched to solve unevenly and insufficiently distributed clinical resources. However, auxiliary diagnosis is still dominated by human physicians, and how to make intelligent systems more involved in the diagnosis process is gradually becoming a concern. An interactive automated clinical diagnosis with a question-answering system and a question generation system can capture a patient’s conditions from multiple perspectives with less physician involvement by asking different questions to drive and guide the diagnosis. This clinical diagnosis process requires diverse information to evaluate a patient from different perspectives to obtain an accurate diagnosis. Recently proposed medical question generation systems have not considered diversity. Thus, we propose a diversity learning-based visual question generation model using a multi-latent space to generate informative question sets from medical images. The proposed method generates various questions by embedding visual and language information in different latent spaces, whose diversity is trained by our newly proposed loss. We have also added control over the categories of generated questions, making the generated questions directional. Furthermore, we use a new metric named similarity to accurately evaluate the proposed model’s performance. The experimental results on the Slake and VQA-RAD datasets demonstrate that the proposed method can generate questions with diverse information. Our model works with an answering model for interactive automated clinical diagnosis and generates datasets to replace the process of annotation that incurs huge labor costs.
format Online
Article
Text
id pubmed-9919063
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher MDPI
record_format MEDLINE/PubMed
spelling pubmed-99190632023-02-12 Diversity Learning Based on Multi-Latent Space for Medical Image Visual Question Generation Zhu, He Togo, Ren Ogawa, Takahiro Haseyama, Miki Sensors (Basel) Article Auxiliary clinical diagnosis has been researched to solve unevenly and insufficiently distributed clinical resources. However, auxiliary diagnosis is still dominated by human physicians, and how to make intelligent systems more involved in the diagnosis process is gradually becoming a concern. An interactive automated clinical diagnosis with a question-answering system and a question generation system can capture a patient’s conditions from multiple perspectives with less physician involvement by asking different questions to drive and guide the diagnosis. This clinical diagnosis process requires diverse information to evaluate a patient from different perspectives to obtain an accurate diagnosis. Recently proposed medical question generation systems have not considered diversity. Thus, we propose a diversity learning-based visual question generation model using a multi-latent space to generate informative question sets from medical images. The proposed method generates various questions by embedding visual and language information in different latent spaces, whose diversity is trained by our newly proposed loss. We have also added control over the categories of generated questions, making the generated questions directional. Furthermore, we use a new metric named similarity to accurately evaluate the proposed model’s performance. The experimental results on the Slake and VQA-RAD datasets demonstrate that the proposed method can generate questions with diverse information. Our model works with an answering model for interactive automated clinical diagnosis and generates datasets to replace the process of annotation that incurs huge labor costs. MDPI 2023-01-17 /pmc/articles/PMC9919063/ /pubmed/36772095 http://dx.doi.org/10.3390/s23031057 Text en © 2023 by the authors. https://creativecommons.org/licenses/by/4.0/Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
spellingShingle Article
Zhu, He
Togo, Ren
Ogawa, Takahiro
Haseyama, Miki
Diversity Learning Based on Multi-Latent Space for Medical Image Visual Question Generation
title Diversity Learning Based on Multi-Latent Space for Medical Image Visual Question Generation
title_full Diversity Learning Based on Multi-Latent Space for Medical Image Visual Question Generation
title_fullStr Diversity Learning Based on Multi-Latent Space for Medical Image Visual Question Generation
title_full_unstemmed Diversity Learning Based on Multi-Latent Space for Medical Image Visual Question Generation
title_short Diversity Learning Based on Multi-Latent Space for Medical Image Visual Question Generation
title_sort diversity learning based on multi-latent space for medical image visual question generation
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9919063/
https://www.ncbi.nlm.nih.gov/pubmed/36772095
http://dx.doi.org/10.3390/s23031057
work_keys_str_mv AT zhuhe diversitylearningbasedonmultilatentspaceformedicalimagevisualquestiongeneration
AT togoren diversitylearningbasedonmultilatentspaceformedicalimagevisualquestiongeneration
AT ogawatakahiro diversitylearningbasedonmultilatentspaceformedicalimagevisualquestiongeneration
AT haseyamamiki diversitylearningbasedonmultilatentspaceformedicalimagevisualquestiongeneration