Cargando…

Encoder-decoder models for chest X-ray report generation perform no better than unconditioned baselines

High quality radiology reporting of chest X-ray images is of core importance for high-quality patient diagnosis and care. Automatically generated reports can assist radiologists by reducing their workload and even may prevent errors. Machine Learning (ML) models for this task take an X-ray image as...

Descripción completa

Detalles Bibliográficos
Autores principales: Babar, Zaheer, van Laarhoven, Twan, Marchiori, Elena
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8629217/
https://www.ncbi.nlm.nih.gov/pubmed/34843509
http://dx.doi.org/10.1371/journal.pone.0259639
_version_ 1784607157508898816
author Babar, Zaheer
van Laarhoven, Twan
Marchiori, Elena
author_facet Babar, Zaheer
van Laarhoven, Twan
Marchiori, Elena
author_sort Babar, Zaheer
collection PubMed
description High quality radiology reporting of chest X-ray images is of core importance for high-quality patient diagnosis and care. Automatically generated reports can assist radiologists by reducing their workload and even may prevent errors. Machine Learning (ML) models for this task take an X-ray image as input and output a sequence of words. In this work, we show that ML models for this task based on the popular encoder-decoder approach, like ‘Show, Attend and Tell’ (SA&T) have similar or worse performance than models that do not use the input image, called unconditioned baseline. An unconditioned model achieved diagnostic accuracy of 0.91 on the IU chest X-ray dataset, and significantly outperformed SA&T (0.877) and other popular ML models (p-value < 0.001). This unconditioned model also outperformed SA&T and similar ML methods on the BLEU-4 and METEOR metrics. Also, an unconditioned version of SA&T obtained by permuting the reports generated from images of the test set, achieved diagnostic accuracy of 0.862, comparable to that of SA&T (p-value ≥ 0.05).
format Online
Article
Text
id pubmed-8629217
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-86292172021-11-30 Encoder-decoder models for chest X-ray report generation perform no better than unconditioned baselines Babar, Zaheer van Laarhoven, Twan Marchiori, Elena PLoS One Research Article High quality radiology reporting of chest X-ray images is of core importance for high-quality patient diagnosis and care. Automatically generated reports can assist radiologists by reducing their workload and even may prevent errors. Machine Learning (ML) models for this task take an X-ray image as input and output a sequence of words. In this work, we show that ML models for this task based on the popular encoder-decoder approach, like ‘Show, Attend and Tell’ (SA&T) have similar or worse performance than models that do not use the input image, called unconditioned baseline. An unconditioned model achieved diagnostic accuracy of 0.91 on the IU chest X-ray dataset, and significantly outperformed SA&T (0.877) and other popular ML models (p-value < 0.001). This unconditioned model also outperformed SA&T and similar ML methods on the BLEU-4 and METEOR metrics. Also, an unconditioned version of SA&T obtained by permuting the reports generated from images of the test set, achieved diagnostic accuracy of 0.862, comparable to that of SA&T (p-value ≥ 0.05). Public Library of Science 2021-11-29 /pmc/articles/PMC8629217/ /pubmed/34843509 http://dx.doi.org/10.1371/journal.pone.0259639 Text en © 2021 Babar et al https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Babar, Zaheer
van Laarhoven, Twan
Marchiori, Elena
Encoder-decoder models for chest X-ray report generation perform no better than unconditioned baselines
title Encoder-decoder models for chest X-ray report generation perform no better than unconditioned baselines
title_full Encoder-decoder models for chest X-ray report generation perform no better than unconditioned baselines
title_fullStr Encoder-decoder models for chest X-ray report generation perform no better than unconditioned baselines
title_full_unstemmed Encoder-decoder models for chest X-ray report generation perform no better than unconditioned baselines
title_short Encoder-decoder models for chest X-ray report generation perform no better than unconditioned baselines
title_sort encoder-decoder models for chest x-ray report generation perform no better than unconditioned baselines
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8629217/
https://www.ncbi.nlm.nih.gov/pubmed/34843509
http://dx.doi.org/10.1371/journal.pone.0259639
work_keys_str_mv AT babarzaheer encoderdecodermodelsforchestxrayreportgenerationperformnobetterthanunconditionedbaselines
AT vanlaarhoventwan encoderdecodermodelsforchestxrayreportgenerationperformnobetterthanunconditionedbaselines
AT marchiorielena encoderdecodermodelsforchestxrayreportgenerationperformnobetterthanunconditionedbaselines