Cargando…

Multi-label annotation of text reports from computed tomography of the chest, abdomen, and pelvis using deep learning

BACKGROUND: There is progress to be made in building artificially intelligent systems to detect abnormalities that are not only accurate but can handle the true breadth of findings that radiologists encounter in body (chest, abdomen, and pelvis) computed tomography (CT). Currently, the major bottlen...

Descripción completa

Detalles Bibliográficos
Autores principales: D’Anniballe, Vincent M., Tushar, Fakrul Islam, Faryna, Khrystyna, Han, Songyue, Mazurowski, Maciej A., Rubin, Geoffrey D., Lo, Joseph Y.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: BioMed Central 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9011942/
https://www.ncbi.nlm.nih.gov/pubmed/35428335
http://dx.doi.org/10.1186/s12911-022-01843-4
_version_ 1784687703065886720
author D’Anniballe, Vincent M.
Tushar, Fakrul Islam
Faryna, Khrystyna
Han, Songyue
Mazurowski, Maciej A.
Rubin, Geoffrey D.
Lo, Joseph Y.
author_facet D’Anniballe, Vincent M.
Tushar, Fakrul Islam
Faryna, Khrystyna
Han, Songyue
Mazurowski, Maciej A.
Rubin, Geoffrey D.
Lo, Joseph Y.
author_sort D’Anniballe, Vincent M.
collection PubMed
description BACKGROUND: There is progress to be made in building artificially intelligent systems to detect abnormalities that are not only accurate but can handle the true breadth of findings that radiologists encounter in body (chest, abdomen, and pelvis) computed tomography (CT). Currently, the major bottleneck for developing multi-disease classifiers is a lack of manually annotated data. The purpose of this work was to develop high throughput multi-label annotators for body CT reports that can be applied across a variety of abnormalities, organs, and disease states thereby mitigating the need for human annotation. METHODS: We used a dictionary approach to develop rule-based algorithms (RBA) for extraction of disease labels from radiology text reports. We targeted three organ systems (lungs/pleura, liver/gallbladder, kidneys/ureters) with four diseases per system based on their prevalence in our dataset. To expand the algorithms beyond pre-defined keywords, attention-guided recurrent neural networks (RNN) were trained using the RBA-extracted labels to classify reports as being positive for one or more diseases or normal for each organ system. Alternative effects on disease classification performance were evaluated using random initialization or pre-trained embedding as well as different sizes of training datasets. The RBA was tested on a subset of 2158 manually labeled reports and performance was reported as accuracy and F-score. The RNN was tested against a test set of 48,758 reports labeled by RBA and performance was reported as area under the receiver operating characteristic curve (AUC), with 95% CIs calculated using the DeLong method. RESULTS: Manual validation of the RBA confirmed 91–99% accuracy across the 15 different labels. Our models extracted disease labels from 261,229 radiology reports of 112,501 unique subjects. Pre-trained models outperformed random initialization across all diseases. As the training dataset size was reduced, performance was robust except for a few diseases with a relatively small number of cases. Pre-trained classification AUCs reached > 0.95 for all four disease outcomes and normality across all three organ systems. CONCLUSIONS: Our label-extracting pipeline was able to encompass a variety of cases and diseases in body CT reports by generalizing beyond strict rules with exceptional accuracy. The method described can be easily adapted to enable automated labeling of hospital-scale medical data sets for training image-based disease classifiers. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1186/s12911-022-01843-4.
format Online
Article
Text
id pubmed-9011942
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher BioMed Central
record_format MEDLINE/PubMed
spelling pubmed-90119422022-04-16 Multi-label annotation of text reports from computed tomography of the chest, abdomen, and pelvis using deep learning D’Anniballe, Vincent M. Tushar, Fakrul Islam Faryna, Khrystyna Han, Songyue Mazurowski, Maciej A. Rubin, Geoffrey D. Lo, Joseph Y. BMC Med Inform Decis Mak Research BACKGROUND: There is progress to be made in building artificially intelligent systems to detect abnormalities that are not only accurate but can handle the true breadth of findings that radiologists encounter in body (chest, abdomen, and pelvis) computed tomography (CT). Currently, the major bottleneck for developing multi-disease classifiers is a lack of manually annotated data. The purpose of this work was to develop high throughput multi-label annotators for body CT reports that can be applied across a variety of abnormalities, organs, and disease states thereby mitigating the need for human annotation. METHODS: We used a dictionary approach to develop rule-based algorithms (RBA) for extraction of disease labels from radiology text reports. We targeted three organ systems (lungs/pleura, liver/gallbladder, kidneys/ureters) with four diseases per system based on their prevalence in our dataset. To expand the algorithms beyond pre-defined keywords, attention-guided recurrent neural networks (RNN) were trained using the RBA-extracted labels to classify reports as being positive for one or more diseases or normal for each organ system. Alternative effects on disease classification performance were evaluated using random initialization or pre-trained embedding as well as different sizes of training datasets. The RBA was tested on a subset of 2158 manually labeled reports and performance was reported as accuracy and F-score. The RNN was tested against a test set of 48,758 reports labeled by RBA and performance was reported as area under the receiver operating characteristic curve (AUC), with 95% CIs calculated using the DeLong method. RESULTS: Manual validation of the RBA confirmed 91–99% accuracy across the 15 different labels. Our models extracted disease labels from 261,229 radiology reports of 112,501 unique subjects. Pre-trained models outperformed random initialization across all diseases. As the training dataset size was reduced, performance was robust except for a few diseases with a relatively small number of cases. Pre-trained classification AUCs reached > 0.95 for all four disease outcomes and normality across all three organ systems. CONCLUSIONS: Our label-extracting pipeline was able to encompass a variety of cases and diseases in body CT reports by generalizing beyond strict rules with exceptional accuracy. The method described can be easily adapted to enable automated labeling of hospital-scale medical data sets for training image-based disease classifiers. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1186/s12911-022-01843-4. BioMed Central 2022-04-15 /pmc/articles/PMC9011942/ /pubmed/35428335 http://dx.doi.org/10.1186/s12911-022-01843-4 Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/ (https://creativecommons.org/publicdomain/zero/1.0/) ) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
spellingShingle Research
D’Anniballe, Vincent M.
Tushar, Fakrul Islam
Faryna, Khrystyna
Han, Songyue
Mazurowski, Maciej A.
Rubin, Geoffrey D.
Lo, Joseph Y.
Multi-label annotation of text reports from computed tomography of the chest, abdomen, and pelvis using deep learning
title Multi-label annotation of text reports from computed tomography of the chest, abdomen, and pelvis using deep learning
title_full Multi-label annotation of text reports from computed tomography of the chest, abdomen, and pelvis using deep learning
title_fullStr Multi-label annotation of text reports from computed tomography of the chest, abdomen, and pelvis using deep learning
title_full_unstemmed Multi-label annotation of text reports from computed tomography of the chest, abdomen, and pelvis using deep learning
title_short Multi-label annotation of text reports from computed tomography of the chest, abdomen, and pelvis using deep learning
title_sort multi-label annotation of text reports from computed tomography of the chest, abdomen, and pelvis using deep learning
topic Research
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9011942/
https://www.ncbi.nlm.nih.gov/pubmed/35428335
http://dx.doi.org/10.1186/s12911-022-01843-4
work_keys_str_mv AT danniballevincentm multilabelannotationoftextreportsfromcomputedtomographyofthechestabdomenandpelvisusingdeeplearning
AT tusharfakrulislam multilabelannotationoftextreportsfromcomputedtomographyofthechestabdomenandpelvisusingdeeplearning
AT farynakhrystyna multilabelannotationoftextreportsfromcomputedtomographyofthechestabdomenandpelvisusingdeeplearning
AT hansongyue multilabelannotationoftextreportsfromcomputedtomographyofthechestabdomenandpelvisusingdeeplearning
AT mazurowskimacieja multilabelannotationoftextreportsfromcomputedtomographyofthechestabdomenandpelvisusingdeeplearning
AT rubingeoffreyd multilabelannotationoftextreportsfromcomputedtomographyofthechestabdomenandpelvisusingdeeplearning
AT lojosephy multilabelannotationoftextreportsfromcomputedtomographyofthechestabdomenandpelvisusingdeeplearning