Cargando…

A multidomain fusion model of radiomics and deep learning to discriminate between PDAC and AIP based on (18)F-FDG PET/CT images

PURPOSE: To explore a multidomain fusion model of radiomics and deep learning features based on (18)F-fluorodeoxyglucose positron emission tomography/computed tomography ((18)F-FDG PET/CT) images to distinguish pancreatic ductal adenocarcinoma (PDAC) and autoimmune pancreatitis (AIP), which could ef...

Descripción completa

Detalles Bibliográficos
Autores principales: Wei, Wenting, Jia, Guorong, Wu, Zhongyi, Wang, Tao, Wang, Heng, Wei, Kezhen, Cheng, Chao, Liu, Zhaobang, Zuo, Changjing
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Springer Nature Singapore 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9676903/
https://www.ncbi.nlm.nih.gov/pubmed/36409398
http://dx.doi.org/10.1007/s11604-022-01363-1
_version_ 1784833693705043968
author Wei, Wenting
Jia, Guorong
Wu, Zhongyi
Wang, Tao
Wang, Heng
Wei, Kezhen
Cheng, Chao
Liu, Zhaobang
Zuo, Changjing
author_facet Wei, Wenting
Jia, Guorong
Wu, Zhongyi
Wang, Tao
Wang, Heng
Wei, Kezhen
Cheng, Chao
Liu, Zhaobang
Zuo, Changjing
author_sort Wei, Wenting
collection PubMed
description PURPOSE: To explore a multidomain fusion model of radiomics and deep learning features based on (18)F-fluorodeoxyglucose positron emission tomography/computed tomography ((18)F-FDG PET/CT) images to distinguish pancreatic ductal adenocarcinoma (PDAC) and autoimmune pancreatitis (AIP), which could effectively improve the accuracy of diseases diagnosis. MATERIALS AND METHODS: This retrospective study included 48 patients with AIP (mean age, 65 ± 12.0 years; range, 37–90 years) and 64 patients with PDAC patients (mean age, 66 ± 11.3 years; range, 32–88 years). Three different methods were discussed to identify PDAC and AIP based on (18)F-FDG PET/CT images, including the radiomics model (RAD_model), the deep learning model (DL_model), and the multidomain fusion model (MF_model). We also compared the classification results of PET/CT, PET, and CT images in these three models. In addition, we explored the attributes of deep learning abstract features by analyzing the correlation between radiomics and deep learning features. Five-fold cross-validation was used to calculate receiver operating characteristic (ROC), area under the roc curve (AUC), accuracy (Acc), sensitivity (Sen), and specificity (Spe) to quantitatively evaluate the performance of different classification models. RESULTS: The experimental results showed that the multidomain fusion model had the best comprehensive performance compared with radiomics and deep learning models, and the AUC, accuracy, sensitivity, specificity were 96.4% (95% CI 95.4–97.3%), 90.1% (95% CI 88.7–91.5%), 87.5% (95% CI 84.3–90.6%), and 93.0% (95% CI 90.3–95.6%), respectively. And our study proved that the multimodal features of PET/CT were superior to using either PET or CT features alone. First-order features of radiomics provided valuable complementary information for the deep learning model. CONCLUSION: The preliminary results of this paper demonstrated that our proposed multidomain fusion model fully exploits the value of radiomics and deep learning features based on (18)F-FDG PET/CT images, which provided competitive accuracy for the discrimination of PDAC and AIP.
format Online
Article
Text
id pubmed-9676903
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Springer Nature Singapore
record_format MEDLINE/PubMed
spelling pubmed-96769032022-11-21 A multidomain fusion model of radiomics and deep learning to discriminate between PDAC and AIP based on (18)F-FDG PET/CT images Wei, Wenting Jia, Guorong Wu, Zhongyi Wang, Tao Wang, Heng Wei, Kezhen Cheng, Chao Liu, Zhaobang Zuo, Changjing Jpn J Radiol Original Article PURPOSE: To explore a multidomain fusion model of radiomics and deep learning features based on (18)F-fluorodeoxyglucose positron emission tomography/computed tomography ((18)F-FDG PET/CT) images to distinguish pancreatic ductal adenocarcinoma (PDAC) and autoimmune pancreatitis (AIP), which could effectively improve the accuracy of diseases diagnosis. MATERIALS AND METHODS: This retrospective study included 48 patients with AIP (mean age, 65 ± 12.0 years; range, 37–90 years) and 64 patients with PDAC patients (mean age, 66 ± 11.3 years; range, 32–88 years). Three different methods were discussed to identify PDAC and AIP based on (18)F-FDG PET/CT images, including the radiomics model (RAD_model), the deep learning model (DL_model), and the multidomain fusion model (MF_model). We also compared the classification results of PET/CT, PET, and CT images in these three models. In addition, we explored the attributes of deep learning abstract features by analyzing the correlation between radiomics and deep learning features. Five-fold cross-validation was used to calculate receiver operating characteristic (ROC), area under the roc curve (AUC), accuracy (Acc), sensitivity (Sen), and specificity (Spe) to quantitatively evaluate the performance of different classification models. RESULTS: The experimental results showed that the multidomain fusion model had the best comprehensive performance compared with radiomics and deep learning models, and the AUC, accuracy, sensitivity, specificity were 96.4% (95% CI 95.4–97.3%), 90.1% (95% CI 88.7–91.5%), 87.5% (95% CI 84.3–90.6%), and 93.0% (95% CI 90.3–95.6%), respectively. And our study proved that the multimodal features of PET/CT were superior to using either PET or CT features alone. First-order features of radiomics provided valuable complementary information for the deep learning model. CONCLUSION: The preliminary results of this paper demonstrated that our proposed multidomain fusion model fully exploits the value of radiomics and deep learning features based on (18)F-FDG PET/CT images, which provided competitive accuracy for the discrimination of PDAC and AIP. Springer Nature Singapore 2022-11-21 2023 /pmc/articles/PMC9676903/ /pubmed/36409398 http://dx.doi.org/10.1007/s11604-022-01363-1 Text en © The Author(s) under exclusive licence to Japan Radiological Society 2022, Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law. This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Original Article
Wei, Wenting
Jia, Guorong
Wu, Zhongyi
Wang, Tao
Wang, Heng
Wei, Kezhen
Cheng, Chao
Liu, Zhaobang
Zuo, Changjing
A multidomain fusion model of radiomics and deep learning to discriminate between PDAC and AIP based on (18)F-FDG PET/CT images
title A multidomain fusion model of radiomics and deep learning to discriminate between PDAC and AIP based on (18)F-FDG PET/CT images
title_full A multidomain fusion model of radiomics and deep learning to discriminate between PDAC and AIP based on (18)F-FDG PET/CT images
title_fullStr A multidomain fusion model of radiomics and deep learning to discriminate between PDAC and AIP based on (18)F-FDG PET/CT images
title_full_unstemmed A multidomain fusion model of radiomics and deep learning to discriminate between PDAC and AIP based on (18)F-FDG PET/CT images
title_short A multidomain fusion model of radiomics and deep learning to discriminate between PDAC and AIP based on (18)F-FDG PET/CT images
title_sort multidomain fusion model of radiomics and deep learning to discriminate between pdac and aip based on (18)f-fdg pet/ct images
topic Original Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9676903/
https://www.ncbi.nlm.nih.gov/pubmed/36409398
http://dx.doi.org/10.1007/s11604-022-01363-1
work_keys_str_mv AT weiwenting amultidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT jiaguorong amultidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT wuzhongyi amultidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT wangtao amultidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT wangheng amultidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT weikezhen amultidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT chengchao amultidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT liuzhaobang amultidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT zuochangjing amultidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT weiwenting multidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT jiaguorong multidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT wuzhongyi multidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT wangtao multidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT wangheng multidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT weikezhen multidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT chengchao multidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT liuzhaobang multidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages
AT zuochangjing multidomainfusionmodelofradiomicsanddeeplearningtodiscriminatebetweenpdacandaipbasedon18ffdgpetctimages