Cargando…

Comparative performances of machine learning algorithms in radiomics and impacting factors

There are no current recommendations on which machine learning (ML) algorithms should be used in radiomics. The objective was to compare performances of ML algorithms in radiomics when applied to different clinical questions to determine whether some strategies could give the best and most stable pe...

Descripción completa

Detalles Bibliográficos
Autores principales: Decoux, Antoine, Duron, Loic, Habert, Paul, Roblot, Victoire, Arsovic, Emina, Chassagnon, Guillaume, Arnoux, Armelle, Fournier, Laure
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Nature Publishing Group UK 2023
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10462640/
https://www.ncbi.nlm.nih.gov/pubmed/37640728
http://dx.doi.org/10.1038/s41598-023-39738-7
_version_ 1785098075996422144
author Decoux, Antoine
Duron, Loic
Habert, Paul
Roblot, Victoire
Arsovic, Emina
Chassagnon, Guillaume
Arnoux, Armelle
Fournier, Laure
author_facet Decoux, Antoine
Duron, Loic
Habert, Paul
Roblot, Victoire
Arsovic, Emina
Chassagnon, Guillaume
Arnoux, Armelle
Fournier, Laure
author_sort Decoux, Antoine
collection PubMed
description There are no current recommendations on which machine learning (ML) algorithms should be used in radiomics. The objective was to compare performances of ML algorithms in radiomics when applied to different clinical questions to determine whether some strategies could give the best and most stable performances regardless of datasets. This study compares the performances of nine feature selection algorithms combined with fourteen binary classification algorithms on ten datasets. These datasets included radiomics features and clinical diagnosis for binary clinical classifications including COVID-19 pneumonia or sarcopenia on CT, head and neck, orbital or uterine lesions on MRI. For each dataset, a train-test split was created. Each of the 126 (9 × 14) combinations of feature selection algorithms and classification algorithms was trained and tuned using a ten-fold cross validation, then AUC was computed. This procedure was repeated three times per dataset. Best overall performances were obtained with JMI and JMIM as feature selection algorithms and random forest and linear regression models as classification algorithms. The choice of the classification algorithm was the factor explaining most of the performance variation (10% of total variance). The choice of the feature selection algorithm explained only 2% of variation, while the train-test split explained 9%.
format Online
Article
Text
id pubmed-10462640
institution National Center for Biotechnology Information
language English
publishDate 2023
publisher Nature Publishing Group UK
record_format MEDLINE/PubMed
spelling pubmed-104626402023-08-30 Comparative performances of machine learning algorithms in radiomics and impacting factors Decoux, Antoine Duron, Loic Habert, Paul Roblot, Victoire Arsovic, Emina Chassagnon, Guillaume Arnoux, Armelle Fournier, Laure Sci Rep Article There are no current recommendations on which machine learning (ML) algorithms should be used in radiomics. The objective was to compare performances of ML algorithms in radiomics when applied to different clinical questions to determine whether some strategies could give the best and most stable performances regardless of datasets. This study compares the performances of nine feature selection algorithms combined with fourteen binary classification algorithms on ten datasets. These datasets included radiomics features and clinical diagnosis for binary clinical classifications including COVID-19 pneumonia or sarcopenia on CT, head and neck, orbital or uterine lesions on MRI. For each dataset, a train-test split was created. Each of the 126 (9 × 14) combinations of feature selection algorithms and classification algorithms was trained and tuned using a ten-fold cross validation, then AUC was computed. This procedure was repeated three times per dataset. Best overall performances were obtained with JMI and JMIM as feature selection algorithms and random forest and linear regression models as classification algorithms. The choice of the classification algorithm was the factor explaining most of the performance variation (10% of total variance). The choice of the feature selection algorithm explained only 2% of variation, while the train-test split explained 9%. Nature Publishing Group UK 2023-08-28 /pmc/articles/PMC10462640/ /pubmed/37640728 http://dx.doi.org/10.1038/s41598-023-39738-7 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) .
spellingShingle Article
Decoux, Antoine
Duron, Loic
Habert, Paul
Roblot, Victoire
Arsovic, Emina
Chassagnon, Guillaume
Arnoux, Armelle
Fournier, Laure
Comparative performances of machine learning algorithms in radiomics and impacting factors
title Comparative performances of machine learning algorithms in radiomics and impacting factors
title_full Comparative performances of machine learning algorithms in radiomics and impacting factors
title_fullStr Comparative performances of machine learning algorithms in radiomics and impacting factors
title_full_unstemmed Comparative performances of machine learning algorithms in radiomics and impacting factors
title_short Comparative performances of machine learning algorithms in radiomics and impacting factors
title_sort comparative performances of machine learning algorithms in radiomics and impacting factors
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10462640/
https://www.ncbi.nlm.nih.gov/pubmed/37640728
http://dx.doi.org/10.1038/s41598-023-39738-7
work_keys_str_mv AT decouxantoine comparativeperformancesofmachinelearningalgorithmsinradiomicsandimpactingfactors
AT duronloic comparativeperformancesofmachinelearningalgorithmsinradiomicsandimpactingfactors
AT habertpaul comparativeperformancesofmachinelearningalgorithmsinradiomicsandimpactingfactors
AT roblotvictoire comparativeperformancesofmachinelearningalgorithmsinradiomicsandimpactingfactors
AT arsovicemina comparativeperformancesofmachinelearningalgorithmsinradiomicsandimpactingfactors
AT chassagnonguillaume comparativeperformancesofmachinelearningalgorithmsinradiomicsandimpactingfactors
AT arnouxarmelle comparativeperformancesofmachinelearningalgorithmsinradiomicsandimpactingfactors
AT fournierlaure comparativeperformancesofmachinelearningalgorithmsinradiomicsandimpactingfactors