Cargando…
Knowledge-enhanced visual-language pre-training on chest radiology images
While multi-modal foundation models pre-trained on large-scale data have been successful in natural language understanding and vision recognition, their use in medical domains is still limited due to the fine-grained nature of medical tasks and the high demand for domain knowledge. To address this c...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Nature Publishing Group UK
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10382552/ https://www.ncbi.nlm.nih.gov/pubmed/37507376 http://dx.doi.org/10.1038/s41467-023-40260-7 |
_version_ | 1785080696871583744 |
---|---|
author | Zhang, Xiaoman Wu, Chaoyi Zhang, Ya Xie, Weidi Wang, Yanfeng |
author_facet | Zhang, Xiaoman Wu, Chaoyi Zhang, Ya Xie, Weidi Wang, Yanfeng |
author_sort | Zhang, Xiaoman |
collection | PubMed |
description | While multi-modal foundation models pre-trained on large-scale data have been successful in natural language understanding and vision recognition, their use in medical domains is still limited due to the fine-grained nature of medical tasks and the high demand for domain knowledge. To address this challenge, we propose an approach called Knowledge-enhanced Auto Diagnosis (KAD) which leverages existing medical domain knowledge to guide vision-language pre-training using paired chest X-rays and radiology reports. We evaluate KAD on four external X-ray datasets and demonstrate that its zero-shot performance is not only comparable to that of fully supervised models but also superior to the average of three expert radiologists for three (out of five) pathologies with statistical significance. Moreover, when few-shot annotation is available, KAD outperforms all existing approaches in fine-tuning settings, demonstrating its potential for application in different clinical scenarios. |
format | Online Article Text |
id | pubmed-10382552 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Nature Publishing Group UK |
record_format | MEDLINE/PubMed |
spelling | pubmed-103825522023-07-30 Knowledge-enhanced visual-language pre-training on chest radiology images Zhang, Xiaoman Wu, Chaoyi Zhang, Ya Xie, Weidi Wang, Yanfeng Nat Commun Article While multi-modal foundation models pre-trained on large-scale data have been successful in natural language understanding and vision recognition, their use in medical domains is still limited due to the fine-grained nature of medical tasks and the high demand for domain knowledge. To address this challenge, we propose an approach called Knowledge-enhanced Auto Diagnosis (KAD) which leverages existing medical domain knowledge to guide vision-language pre-training using paired chest X-rays and radiology reports. We evaluate KAD on four external X-ray datasets and demonstrate that its zero-shot performance is not only comparable to that of fully supervised models but also superior to the average of three expert radiologists for three (out of five) pathologies with statistical significance. Moreover, when few-shot annotation is available, KAD outperforms all existing approaches in fine-tuning settings, demonstrating its potential for application in different clinical scenarios. Nature Publishing Group UK 2023-07-28 /pmc/articles/PMC10382552/ /pubmed/37507376 http://dx.doi.org/10.1038/s41467-023-40260-7 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Article Zhang, Xiaoman Wu, Chaoyi Zhang, Ya Xie, Weidi Wang, Yanfeng Knowledge-enhanced visual-language pre-training on chest radiology images |
title | Knowledge-enhanced visual-language pre-training on chest radiology images |
title_full | Knowledge-enhanced visual-language pre-training on chest radiology images |
title_fullStr | Knowledge-enhanced visual-language pre-training on chest radiology images |
title_full_unstemmed | Knowledge-enhanced visual-language pre-training on chest radiology images |
title_short | Knowledge-enhanced visual-language pre-training on chest radiology images |
title_sort | knowledge-enhanced visual-language pre-training on chest radiology images |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10382552/ https://www.ncbi.nlm.nih.gov/pubmed/37507376 http://dx.doi.org/10.1038/s41467-023-40260-7 |
work_keys_str_mv | AT zhangxiaoman knowledgeenhancedvisuallanguagepretrainingonchestradiologyimages AT wuchaoyi knowledgeenhancedvisuallanguagepretrainingonchestradiologyimages AT zhangya knowledgeenhancedvisuallanguagepretrainingonchestradiologyimages AT xieweidi knowledgeenhancedvisuallanguagepretrainingonchestradiologyimages AT wangyanfeng knowledgeenhancedvisuallanguagepretrainingonchestradiologyimages |