Cargando…

RadImageNet: An Open Radiologic Deep Learning Research Dataset for Effective Transfer Learning

PURPOSE: To demonstrate the value of pretraining with millions of radiologic images compared with ImageNet photographic images on downstream medical applications when using transfer learning. MATERIALS AND METHODS: This retrospective study included patients who underwent a radiologic study between 2...

Descripción completa

Detalles Bibliográficos
Autores principales: Mei, Xueyan, Liu, Zelong, Robson, Philip M., Marinelli, Brett, Huang, Mingqian, Doshi, Amish, Jacobi, Adam, Cao, Chendi, Link, Katherine E., Yang, Thomas, Wang, Ying, Greenspan, Hayit, Deyer, Timothy, Fayad, Zahi A., Yang, Yang
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Radiological Society of North America 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9530758/
https://www.ncbi.nlm.nih.gov/pubmed/36204533
http://dx.doi.org/10.1148/ryai.210315
_version_ 1784801755262877696
author Mei, Xueyan
Liu, Zelong
Robson, Philip M.
Marinelli, Brett
Huang, Mingqian
Doshi, Amish
Jacobi, Adam
Cao, Chendi
Link, Katherine E.
Yang, Thomas
Wang, Ying
Greenspan, Hayit
Deyer, Timothy
Fayad, Zahi A.
Yang, Yang
author_facet Mei, Xueyan
Liu, Zelong
Robson, Philip M.
Marinelli, Brett
Huang, Mingqian
Doshi, Amish
Jacobi, Adam
Cao, Chendi
Link, Katherine E.
Yang, Thomas
Wang, Ying
Greenspan, Hayit
Deyer, Timothy
Fayad, Zahi A.
Yang, Yang
author_sort Mei, Xueyan
collection PubMed
description PURPOSE: To demonstrate the value of pretraining with millions of radiologic images compared with ImageNet photographic images on downstream medical applications when using transfer learning. MATERIALS AND METHODS: This retrospective study included patients who underwent a radiologic study between 2005 and 2020 at an outpatient imaging facility. Key images and associated labels from the studies were retrospectively extracted from the original study interpretation. These images were used for RadImageNet model training with random weight initiation. The RadImageNet models were compared with ImageNet models using the area under the receiver operating characteristic curve (AUC) for eight classification tasks and using Dice scores for two segmentation problems. RESULTS: The RadImageNet database consists of 1.35 million annotated medical images in 131 872 patients who underwent CT, MRI, and US for musculoskeletal, neurologic, oncologic, gastrointestinal, endocrine, abdominal, and pulmonary pathologic conditions. For transfer learning tasks on small datasets—thyroid nodules (US), breast masses (US), anterior cruciate ligament injuries (MRI), and meniscal tears (MRI)—the RadImageNet models demonstrated a significant advantage (P < .001) to ImageNet models (9.4%, 4.0%, 4.8%, and 4.5% AUC improvements, respectively). For larger datasets—pneumonia (chest radiography), COVID-19 (CT), SARS-CoV-2 (CT), and intracranial hemorrhage (CT)—the RadImageNet models also illustrated improved AUC (P < .001) by 1.9%, 6.1%, 1.7%, and 0.9%, respectively. Additionally, lesion localizations of the RadImageNet models were improved by 64.6% and 16.4% on thyroid and breast US datasets, respectively. CONCLUSION: RadImageNet pretrained models demonstrated better interpretability compared with ImageNet models, especially for smaller radiologic datasets. Keywords: CT, MR Imaging, US, Head/Neck, Thorax, Brain/Brain Stem, Evidence-based Medicine, Computer Applications–General (Informatics) Supplemental material is available for this article. Published under a CC BY 4.0 license. See also the commentary by Cadrin-Chênevert in this issue.
format Online
Article
Text
id pubmed-9530758
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Radiological Society of North America
record_format MEDLINE/PubMed
spelling pubmed-95307582022-10-05 RadImageNet: An Open Radiologic Deep Learning Research Dataset for Effective Transfer Learning Mei, Xueyan Liu, Zelong Robson, Philip M. Marinelli, Brett Huang, Mingqian Doshi, Amish Jacobi, Adam Cao, Chendi Link, Katherine E. Yang, Thomas Wang, Ying Greenspan, Hayit Deyer, Timothy Fayad, Zahi A. Yang, Yang Radiol Artif Intell Original Research PURPOSE: To demonstrate the value of pretraining with millions of radiologic images compared with ImageNet photographic images on downstream medical applications when using transfer learning. MATERIALS AND METHODS: This retrospective study included patients who underwent a radiologic study between 2005 and 2020 at an outpatient imaging facility. Key images and associated labels from the studies were retrospectively extracted from the original study interpretation. These images were used for RadImageNet model training with random weight initiation. The RadImageNet models were compared with ImageNet models using the area under the receiver operating characteristic curve (AUC) for eight classification tasks and using Dice scores for two segmentation problems. RESULTS: The RadImageNet database consists of 1.35 million annotated medical images in 131 872 patients who underwent CT, MRI, and US for musculoskeletal, neurologic, oncologic, gastrointestinal, endocrine, abdominal, and pulmonary pathologic conditions. For transfer learning tasks on small datasets—thyroid nodules (US), breast masses (US), anterior cruciate ligament injuries (MRI), and meniscal tears (MRI)—the RadImageNet models demonstrated a significant advantage (P < .001) to ImageNet models (9.4%, 4.0%, 4.8%, and 4.5% AUC improvements, respectively). For larger datasets—pneumonia (chest radiography), COVID-19 (CT), SARS-CoV-2 (CT), and intracranial hemorrhage (CT)—the RadImageNet models also illustrated improved AUC (P < .001) by 1.9%, 6.1%, 1.7%, and 0.9%, respectively. Additionally, lesion localizations of the RadImageNet models were improved by 64.6% and 16.4% on thyroid and breast US datasets, respectively. CONCLUSION: RadImageNet pretrained models demonstrated better interpretability compared with ImageNet models, especially for smaller radiologic datasets. Keywords: CT, MR Imaging, US, Head/Neck, Thorax, Brain/Brain Stem, Evidence-based Medicine, Computer Applications–General (Informatics) Supplemental material is available for this article. Published under a CC BY 4.0 license. See also the commentary by Cadrin-Chênevert in this issue. Radiological Society of North America 2022-07-27 /pmc/articles/PMC9530758/ /pubmed/36204533 http://dx.doi.org/10.1148/ryai.210315 Text en © 2022 by the Radiological Society of North America, Inc. https://creativecommons.org/licenses/by/4.0/Published under a (https://creativecommons.org/licenses/by/4.0/) CC BY 4.0 license.
spellingShingle Original Research
Mei, Xueyan
Liu, Zelong
Robson, Philip M.
Marinelli, Brett
Huang, Mingqian
Doshi, Amish
Jacobi, Adam
Cao, Chendi
Link, Katherine E.
Yang, Thomas
Wang, Ying
Greenspan, Hayit
Deyer, Timothy
Fayad, Zahi A.
Yang, Yang
RadImageNet: An Open Radiologic Deep Learning Research Dataset for Effective Transfer Learning
title RadImageNet: An Open Radiologic Deep Learning Research Dataset for Effective Transfer Learning
title_full RadImageNet: An Open Radiologic Deep Learning Research Dataset for Effective Transfer Learning
title_fullStr RadImageNet: An Open Radiologic Deep Learning Research Dataset for Effective Transfer Learning
title_full_unstemmed RadImageNet: An Open Radiologic Deep Learning Research Dataset for Effective Transfer Learning
title_short RadImageNet: An Open Radiologic Deep Learning Research Dataset for Effective Transfer Learning
title_sort radimagenet: an open radiologic deep learning research dataset for effective transfer learning
topic Original Research
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9530758/
https://www.ncbi.nlm.nih.gov/pubmed/36204533
http://dx.doi.org/10.1148/ryai.210315
work_keys_str_mv AT meixueyan radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT liuzelong radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT robsonphilipm radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT marinellibrett radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT huangmingqian radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT doshiamish radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT jacobiadam radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT caochendi radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT linkkatherinee radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT yangthomas radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT wangying radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT greenspanhayit radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT deyertimothy radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT fayadzahia radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning
AT yangyang radimagenetanopenradiologicdeeplearningresearchdatasetforeffectivetransferlearning