Cargando…
Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents
IMPORTANCE: Chest radiography is the most common diagnostic imaging examination performed in emergency departments (EDs). Augmenting clinicians with automated preliminary read assistants could help expedite their workflows, improve accuracy, and reduce the cost of care. OBJECTIVE: To assess the perf...
Autores principales: | , , , , , , , , , , , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
American Medical Association
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7547369/ https://www.ncbi.nlm.nih.gov/pubmed/33034642 http://dx.doi.org/10.1001/jamanetworkopen.2020.22779 |
_version_ | 1783592405813952512 |
---|---|
author | Wu, Joy T. Wong, Ken C. L. Gur, Yaniv Ansari, Nadeem Karargyris, Alexandros Sharma, Arjun Morris, Michael Saboury, Babak Ahmad, Hassan Boyko, Orest Syed, Ali Jadhav, Ashutosh Wang, Hongzhi Pillai, Anup Kashyap, Satyananda Moradi, Mehdi Syeda-Mahmood, Tanveer |
author_facet | Wu, Joy T. Wong, Ken C. L. Gur, Yaniv Ansari, Nadeem Karargyris, Alexandros Sharma, Arjun Morris, Michael Saboury, Babak Ahmad, Hassan Boyko, Orest Syed, Ali Jadhav, Ashutosh Wang, Hongzhi Pillai, Anup Kashyap, Satyananda Moradi, Mehdi Syeda-Mahmood, Tanveer |
author_sort | Wu, Joy T. |
collection | PubMed |
description | IMPORTANCE: Chest radiography is the most common diagnostic imaging examination performed in emergency departments (EDs). Augmenting clinicians with automated preliminary read assistants could help expedite their workflows, improve accuracy, and reduce the cost of care. OBJECTIVE: To assess the performance of artificial intelligence (AI) algorithms in realistic radiology workflows by performing an objective comparative evaluation of the preliminary reads of anteroposterior (AP) frontal chest radiographs performed by an AI algorithm and radiology residents. DESIGN, SETTING, AND PARTICIPANTS: This diagnostic study included a set of 72 findings assembled by clinical experts to constitute a full-fledged preliminary read of AP frontal chest radiographs. A novel deep learning architecture was designed for an AI algorithm to estimate the findings per image. The AI algorithm was trained using a multihospital training data set of 342 126 frontal chest radiographs captured in ED and urgent care settings. The training data were labeled from their associated reports. Image-based F1 score was chosen to optimize the operating point on the receiver operating characteristics (ROC) curve so as to minimize the number of missed findings and overcalls per image read. The performance of the model was compared with that of 5 radiology residents recruited from multiple institutions in the US in an objective study in which a separate data set of 1998 AP frontal chest radiographs was drawn from a hospital source representative of realistic preliminary reads in inpatient and ED settings. A triple consensus with adjudication process was used to derive the ground truth labels for the study data set. The performance of AI algorithm and radiology residents was assessed by comparing their reads with ground truth findings. All studies were conducted through a web-based clinical study application system. The triple consensus data set was collected between February and October 2018. The comparison study was preformed between January and October 2019. Data were analyzed from October to February 2020. After the first round of reviews, further analysis of the data was performed from March to July 2020. MAIN OUTCOMES AND MEASURES: The learning performance of the AI algorithm was judged using the conventional ROC curve and the area under the curve (AUC) during training and field testing on the study data set. For the AI algorithm and radiology residents, the individual finding label performance was measured using the conventional measures of label-based sensitivity, specificity, and positive predictive value (PPV). In addition, the agreement with the ground truth on the assignment of findings to images was measured using the pooled κ statistic. The preliminary read performance was recorded for AI algorithm and radiology residents using new measures of mean image-based sensitivity, specificity, and PPV designed for recording the fraction of misses and overcalls on a per image basis. The 1-sided analysis of variance test was used to compare the means of each group (AI algorithm vs radiology residents) using the F distribution, and the null hypothesis was that the groups would have similar means. RESULTS: The trained AI algorithm achieved a mean AUC across labels of 0.807 (weighted mean AUC, 0.841) after training. On the study data set, which had a different prevalence distribution, the mean AUC achieved was 0.772 (weighted mean AUC, 0.865). The interrater agreement with ground truth finding labels for AI algorithm predictions had pooled κ value of 0.544, and the pooled κ for radiology residents was 0.585. For the preliminary read performance, the analysis of variance test was used to compare the distributions of AI algorithm and radiology residents’ mean image-based sensitivity, PPV, and specificity. The mean image-based sensitivity for AI algorithm was 0.716 (95% CI, 0.704-0.729) and for radiology residents was 0.720 (95% CI, 0.709-0.732) (P = .66), while the PPV was 0.730 (95% CI, 0.718-0.742) for the AI algorithm and 0.682 (95% CI, 0.670-0.694) for the radiology residents (P < .001), and specificity was 0.980 (95% CI, 0.980-0.981) for the AI algorithm and 0.973 (95% CI, 0.971-0.974) for the radiology residents (P < .001). CONCLUSIONS AND RELEVANCE: These findings suggest that it is possible to build AI algorithms that reach and exceed the mean level of performance of third-year radiology residents for full-fledged preliminary read of AP frontal chest radiographs. This diagnostic study also found that while the more complex findings would still benefit from expert overreads, the performance of AI algorithms was associated with the amount of data available for training rather than the level of difficulty of interpretation of the finding. Integrating such AI systems in radiology workflows for preliminary interpretations has the potential to expedite existing radiology workflows and address resource scarcity while improving overall accuracy and reducing the cost of care. |
format | Online Article Text |
id | pubmed-7547369 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | American Medical Association |
record_format | MEDLINE/PubMed |
spelling | pubmed-75473692020-10-20 Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents Wu, Joy T. Wong, Ken C. L. Gur, Yaniv Ansari, Nadeem Karargyris, Alexandros Sharma, Arjun Morris, Michael Saboury, Babak Ahmad, Hassan Boyko, Orest Syed, Ali Jadhav, Ashutosh Wang, Hongzhi Pillai, Anup Kashyap, Satyananda Moradi, Mehdi Syeda-Mahmood, Tanveer JAMA Netw Open Original Investigation IMPORTANCE: Chest radiography is the most common diagnostic imaging examination performed in emergency departments (EDs). Augmenting clinicians with automated preliminary read assistants could help expedite their workflows, improve accuracy, and reduce the cost of care. OBJECTIVE: To assess the performance of artificial intelligence (AI) algorithms in realistic radiology workflows by performing an objective comparative evaluation of the preliminary reads of anteroposterior (AP) frontal chest radiographs performed by an AI algorithm and radiology residents. DESIGN, SETTING, AND PARTICIPANTS: This diagnostic study included a set of 72 findings assembled by clinical experts to constitute a full-fledged preliminary read of AP frontal chest radiographs. A novel deep learning architecture was designed for an AI algorithm to estimate the findings per image. The AI algorithm was trained using a multihospital training data set of 342 126 frontal chest radiographs captured in ED and urgent care settings. The training data were labeled from their associated reports. Image-based F1 score was chosen to optimize the operating point on the receiver operating characteristics (ROC) curve so as to minimize the number of missed findings and overcalls per image read. The performance of the model was compared with that of 5 radiology residents recruited from multiple institutions in the US in an objective study in which a separate data set of 1998 AP frontal chest radiographs was drawn from a hospital source representative of realistic preliminary reads in inpatient and ED settings. A triple consensus with adjudication process was used to derive the ground truth labels for the study data set. The performance of AI algorithm and radiology residents was assessed by comparing their reads with ground truth findings. All studies were conducted through a web-based clinical study application system. The triple consensus data set was collected between February and October 2018. The comparison study was preformed between January and October 2019. Data were analyzed from October to February 2020. After the first round of reviews, further analysis of the data was performed from March to July 2020. MAIN OUTCOMES AND MEASURES: The learning performance of the AI algorithm was judged using the conventional ROC curve and the area under the curve (AUC) during training and field testing on the study data set. For the AI algorithm and radiology residents, the individual finding label performance was measured using the conventional measures of label-based sensitivity, specificity, and positive predictive value (PPV). In addition, the agreement with the ground truth on the assignment of findings to images was measured using the pooled κ statistic. The preliminary read performance was recorded for AI algorithm and radiology residents using new measures of mean image-based sensitivity, specificity, and PPV designed for recording the fraction of misses and overcalls on a per image basis. The 1-sided analysis of variance test was used to compare the means of each group (AI algorithm vs radiology residents) using the F distribution, and the null hypothesis was that the groups would have similar means. RESULTS: The trained AI algorithm achieved a mean AUC across labels of 0.807 (weighted mean AUC, 0.841) after training. On the study data set, which had a different prevalence distribution, the mean AUC achieved was 0.772 (weighted mean AUC, 0.865). The interrater agreement with ground truth finding labels for AI algorithm predictions had pooled κ value of 0.544, and the pooled κ for radiology residents was 0.585. For the preliminary read performance, the analysis of variance test was used to compare the distributions of AI algorithm and radiology residents’ mean image-based sensitivity, PPV, and specificity. The mean image-based sensitivity for AI algorithm was 0.716 (95% CI, 0.704-0.729) and for radiology residents was 0.720 (95% CI, 0.709-0.732) (P = .66), while the PPV was 0.730 (95% CI, 0.718-0.742) for the AI algorithm and 0.682 (95% CI, 0.670-0.694) for the radiology residents (P < .001), and specificity was 0.980 (95% CI, 0.980-0.981) for the AI algorithm and 0.973 (95% CI, 0.971-0.974) for the radiology residents (P < .001). CONCLUSIONS AND RELEVANCE: These findings suggest that it is possible to build AI algorithms that reach and exceed the mean level of performance of third-year radiology residents for full-fledged preliminary read of AP frontal chest radiographs. This diagnostic study also found that while the more complex findings would still benefit from expert overreads, the performance of AI algorithms was associated with the amount of data available for training rather than the level of difficulty of interpretation of the finding. Integrating such AI systems in radiology workflows for preliminary interpretations has the potential to expedite existing radiology workflows and address resource scarcity while improving overall accuracy and reducing the cost of care. American Medical Association 2020-10-09 /pmc/articles/PMC7547369/ /pubmed/33034642 http://dx.doi.org/10.1001/jamanetworkopen.2020.22779 Text en Copyright 2020 Wu JT et al. JAMA Network Open. https://creativecommons.org/licenses/by-nc-nd/4.0/This is an open access article distributed under the terms of the CC-BY-NC-ND License. |
spellingShingle | Original Investigation Wu, Joy T. Wong, Ken C. L. Gur, Yaniv Ansari, Nadeem Karargyris, Alexandros Sharma, Arjun Morris, Michael Saboury, Babak Ahmad, Hassan Boyko, Orest Syed, Ali Jadhav, Ashutosh Wang, Hongzhi Pillai, Anup Kashyap, Satyananda Moradi, Mehdi Syeda-Mahmood, Tanveer Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents |
title | Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents |
title_full | Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents |
title_fullStr | Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents |
title_full_unstemmed | Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents |
title_short | Comparison of Chest Radiograph Interpretations by Artificial Intelligence Algorithm vs Radiology Residents |
title_sort | comparison of chest radiograph interpretations by artificial intelligence algorithm vs radiology residents |
topic | Original Investigation |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7547369/ https://www.ncbi.nlm.nih.gov/pubmed/33034642 http://dx.doi.org/10.1001/jamanetworkopen.2020.22779 |
work_keys_str_mv | AT wujoyt comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT wongkencl comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT guryaniv comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT ansarinadeem comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT karargyrisalexandros comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT sharmaarjun comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT morrismichael comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT sabourybabak comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT ahmadhassan comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT boykoorest comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT syedali comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT jadhavashutosh comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT wanghongzhi comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT pillaianup comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT kashyapsatyananda comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT moradimehdi comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents AT syedamahmoodtanveer comparisonofchestradiographinterpretationsbyartificialintelligencealgorithmvsradiologyresidents |