Cargando…

Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study

Introduction: Prognostic scores are important tools in oncology to facilitate clinical decision-making based on patient characteristics. To date, classic survival analysis using Cox proportional hazards regression has been employed in the development of these prognostic scores. With the advance of a...

Descripción completa

Detalles Bibliográficos
Autores principales: Loureiro, Hugo, Becker, Tim, Bauer-Mehren, Anna, Ahmidi, Narges, Weberpals, Janick
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2021
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8086599/
https://www.ncbi.nlm.nih.gov/pubmed/33937744
http://dx.doi.org/10.3389/frai.2021.625573
_version_ 1783686539787632640
author Loureiro, Hugo
Becker, Tim
Bauer-Mehren, Anna
Ahmidi, Narges
Weberpals, Janick
author_facet Loureiro, Hugo
Becker, Tim
Bauer-Mehren, Anna
Ahmidi, Narges
Weberpals, Janick
author_sort Loureiro, Hugo
collection PubMed
description Introduction: Prognostic scores are important tools in oncology to facilitate clinical decision-making based on patient characteristics. To date, classic survival analysis using Cox proportional hazards regression has been employed in the development of these prognostic scores. With the advance of analytical models, this study aimed to determine if more complex machine-learning algorithms could outperform classical survival analysis methods. Methods: In this benchmarking study, two datasets were used to develop and compare different prognostic models for overall survival in pan-cancer populations: a nationwide EHR-derived de-identified database for training and in-sample testing and the OAK (phase III clinical trial) dataset for out-of-sample testing. A real-world database comprised 136K first-line treated cancer patients across multiple cancer types and was split into a 90% training and 10% testing dataset, respectively. The OAK dataset comprised 1,187 patients diagnosed with non-small cell lung cancer. To assess the effect of the covariate number on prognostic performance, we formed three feature sets with 27, 44 and 88 covariates. In terms of methods, we benchmarked ROPRO, a prognostic score based on the Cox model, against eight complex machine-learning models: regularized Cox, Random Survival Forests (RSF), Gradient Boosting (GB), DeepSurv (DS), Autoencoder (AE) and Super Learner (SL). The C-index was used as the performance metric to compare different models. Results: For in-sample testing on the real-world database the resulting C-index [95% CI] values for RSF 0.720 [0.716, 0.725], GB 0.722 [0.718, 0.727], DS 0.721 [0.717, 0.726] and lastly, SL 0.723 [0.718, 0.728] showed significantly better performance as compared to ROPRO 0.701 [0.696, 0.706]. Similar results were derived across all feature sets. However, for the out-of-sample validation on OAK, the stronger performance of the more complex models was not apparent anymore. Consistently, the increase in the number of prognostic covariates did not lead to an increase in model performance. Discussion: The stronger performance of the more complex models did not generalize when applied to an out-of-sample dataset. We hypothesize that future research may benefit by adding multimodal data to exploit advantages of more complex models.
format Online
Article
Text
id pubmed-8086599
institution National Center for Biotechnology Information
language English
publishDate 2021
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-80865992021-05-01 Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study Loureiro, Hugo Becker, Tim Bauer-Mehren, Anna Ahmidi, Narges Weberpals, Janick Front Artif Intell Artificial Intelligence Introduction: Prognostic scores are important tools in oncology to facilitate clinical decision-making based on patient characteristics. To date, classic survival analysis using Cox proportional hazards regression has been employed in the development of these prognostic scores. With the advance of analytical models, this study aimed to determine if more complex machine-learning algorithms could outperform classical survival analysis methods. Methods: In this benchmarking study, two datasets were used to develop and compare different prognostic models for overall survival in pan-cancer populations: a nationwide EHR-derived de-identified database for training and in-sample testing and the OAK (phase III clinical trial) dataset for out-of-sample testing. A real-world database comprised 136K first-line treated cancer patients across multiple cancer types and was split into a 90% training and 10% testing dataset, respectively. The OAK dataset comprised 1,187 patients diagnosed with non-small cell lung cancer. To assess the effect of the covariate number on prognostic performance, we formed three feature sets with 27, 44 and 88 covariates. In terms of methods, we benchmarked ROPRO, a prognostic score based on the Cox model, against eight complex machine-learning models: regularized Cox, Random Survival Forests (RSF), Gradient Boosting (GB), DeepSurv (DS), Autoencoder (AE) and Super Learner (SL). The C-index was used as the performance metric to compare different models. Results: For in-sample testing on the real-world database the resulting C-index [95% CI] values for RSF 0.720 [0.716, 0.725], GB 0.722 [0.718, 0.727], DS 0.721 [0.717, 0.726] and lastly, SL 0.723 [0.718, 0.728] showed significantly better performance as compared to ROPRO 0.701 [0.696, 0.706]. Similar results were derived across all feature sets. However, for the out-of-sample validation on OAK, the stronger performance of the more complex models was not apparent anymore. Consistently, the increase in the number of prognostic covariates did not lead to an increase in model performance. Discussion: The stronger performance of the more complex models did not generalize when applied to an out-of-sample dataset. We hypothesize that future research may benefit by adding multimodal data to exploit advantages of more complex models. Frontiers Media S.A. 2021-04-16 /pmc/articles/PMC8086599/ /pubmed/33937744 http://dx.doi.org/10.3389/frai.2021.625573 Text en Copyright © 2021 Loureiro, Becker, Bauer-Mehren, Ahmidi and Weberpals. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Loureiro, Hugo
Becker, Tim
Bauer-Mehren, Anna
Ahmidi, Narges
Weberpals, Janick
Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study
title Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study
title_full Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study
title_fullStr Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study
title_full_unstemmed Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study
title_short Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study
title_sort artificial intelligence for prognostic scores in oncology: a benchmarking study
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8086599/
https://www.ncbi.nlm.nih.gov/pubmed/33937744
http://dx.doi.org/10.3389/frai.2021.625573
work_keys_str_mv AT loureirohugo artificialintelligenceforprognosticscoresinoncologyabenchmarkingstudy
AT beckertim artificialintelligenceforprognosticscoresinoncologyabenchmarkingstudy
AT bauermehrenanna artificialintelligenceforprognosticscoresinoncologyabenchmarkingstudy
AT ahmidinarges artificialintelligenceforprognosticscoresinoncologyabenchmarkingstudy
AT weberpalsjanick artificialintelligenceforprognosticscoresinoncologyabenchmarkingstudy