Cargando…

Understanding transfer learning for chest radiograph clinical report generation with modified transformer architectures

The image captioning task is increasingly prevalent in artificial intelligence applications for medicine. One important application is clinical report generation from chest radiographs. The clinical writing of unstructured reports is time consuming and error-prone. An automated system would improve...

Descripción completa

Detalles Bibliográficos
Autores principales: Vendrow, Edward, Schonfeld, Ethan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Elsevier 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10372225/
https://www.ncbi.nlm.nih.gov/pubmed/37519756
http://dx.doi.org/10.1016/j.heliyon.2023.e17968
_version_ 1785078325354430464
author Vendrow, Edward
Schonfeld, Ethan
author_facet Vendrow, Edward
Schonfeld, Ethan
author_sort Vendrow, Edward
collection PubMed
description The image captioning task is increasingly prevalent in artificial intelligence applications for medicine. One important application is clinical report generation from chest radiographs. The clinical writing of unstructured reports is time consuming and error-prone. An automated system would improve standardization, error reduction, time consumption, and medical accessibility. In this paper we demonstrate the importance of domain specific pre-training and propose a modified transformer architecture for the medical image captioning task. To accomplish this, we train a series of modified transformers to generate clinical reports from chest radiograph image input. These modified transformers include: a meshed-memory augmented transformer architecture with visual extractor using ImageNet pre-trained weights, a meshed-memory augmented transformer architecture with visual extractor using CheXpert pre-trained weights, and a meshed-memory augmented transformer whose encoder is passed the concatenated embeddings using both ImageNet pre-trained weights and CheXpert pre-trained weights. We use BLEU(1-4), ROUGE-L, CIDEr, and the clinical CheXbert F1 scores to validate our models and demonstrate competitive scores with state of the art models. We provide evidence that ImageNet pre-training is ill-suited for the medical image captioning task, especially for less frequent conditions (e.g.: enlarged cardiomediastinum, lung lesion, pneumothorax). Furthermore, we demonstrate that the double feature model improves performance for specific medical conditions (edema, consolidation, pneumothorax, support devices) and overall CheXbert F1 score, and should be further developed in future work. Such a double feature model, including both ImageNet pre-training as well as domain specific pre-training, could be used in a wide range of image captioning models in medicine.
format Online
Article
Text
id pubmed-10372225
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Elsevier
record_format MEDLINE/PubMed
spelling pubmed-103722252023-07-28 Understanding transfer learning for chest radiograph clinical report generation with modified transformer architectures Vendrow, Edward Schonfeld, Ethan Heliyon Research Article The image captioning task is increasingly prevalent in artificial intelligence applications for medicine. One important application is clinical report generation from chest radiographs. The clinical writing of unstructured reports is time consuming and error-prone. An automated system would improve standardization, error reduction, time consumption, and medical accessibility. In this paper we demonstrate the importance of domain specific pre-training and propose a modified transformer architecture for the medical image captioning task. To accomplish this, we train a series of modified transformers to generate clinical reports from chest radiograph image input. These modified transformers include: a meshed-memory augmented transformer architecture with visual extractor using ImageNet pre-trained weights, a meshed-memory augmented transformer architecture with visual extractor using CheXpert pre-trained weights, and a meshed-memory augmented transformer whose encoder is passed the concatenated embeddings using both ImageNet pre-trained weights and CheXpert pre-trained weights. We use BLEU(1-4), ROUGE-L, CIDEr, and the clinical CheXbert F1 scores to validate our models and demonstrate competitive scores with state of the art models. We provide evidence that ImageNet pre-training is ill-suited for the medical image captioning task, especially for less frequent conditions (e.g.: enlarged cardiomediastinum, lung lesion, pneumothorax). Furthermore, we demonstrate that the double feature model improves performance for specific medical conditions (edema, consolidation, pneumothorax, support devices) and overall CheXbert F1 score, and should be further developed in future work. Such a double feature model, including both ImageNet pre-training as well as domain specific pre-training, could be used in a wide range of image captioning models in medicine. Elsevier 2023-07-10 /pmc/articles/PMC10372225/ /pubmed/37519756 http://dx.doi.org/10.1016/j.heliyon.2023.e17968 Text en © 2023 The Authors. Published by Elsevier Ltd. https://creativecommons.org/licenses/by/4.0/This is an open access article under the CC BY license (http://creativecommons.org/licenses/by/4.0/).
spellingShingle Research Article
Vendrow, Edward
Schonfeld, Ethan
Understanding transfer learning for chest radiograph clinical report generation with modified transformer architectures
title Understanding transfer learning for chest radiograph clinical report generation with modified transformer architectures
title_full Understanding transfer learning for chest radiograph clinical report generation with modified transformer architectures
title_fullStr Understanding transfer learning for chest radiograph clinical report generation with modified transformer architectures
title_full_unstemmed Understanding transfer learning for chest radiograph clinical report generation with modified transformer architectures
title_short Understanding transfer learning for chest radiograph clinical report generation with modified transformer architectures
title_sort understanding transfer learning for chest radiograph clinical report generation with modified transformer architectures
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10372225/
https://www.ncbi.nlm.nih.gov/pubmed/37519756
http://dx.doi.org/10.1016/j.heliyon.2023.e17968
work_keys_str_mv AT vendrowedward understandingtransferlearningforchestradiographclinicalreportgenerationwithmodifiedtransformerarchitectures
AT schonfeldethan understandingtransferlearningforchestradiographclinicalreportgenerationwithmodifiedtransformerarchitectures