Cargando…
Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study
Introduction: Prognostic scores are important tools in oncology to facilitate clinical decision-making based on patient characteristics. To date, classic survival analysis using Cox proportional hazards regression has been employed in the development of these prognostic scores. With the advance of a...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2021
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8086599/ https://www.ncbi.nlm.nih.gov/pubmed/33937744 http://dx.doi.org/10.3389/frai.2021.625573 |
_version_ | 1783686539787632640 |
---|---|
author | Loureiro, Hugo Becker, Tim Bauer-Mehren, Anna Ahmidi, Narges Weberpals, Janick |
author_facet | Loureiro, Hugo Becker, Tim Bauer-Mehren, Anna Ahmidi, Narges Weberpals, Janick |
author_sort | Loureiro, Hugo |
collection | PubMed |
description | Introduction: Prognostic scores are important tools in oncology to facilitate clinical decision-making based on patient characteristics. To date, classic survival analysis using Cox proportional hazards regression has been employed in the development of these prognostic scores. With the advance of analytical models, this study aimed to determine if more complex machine-learning algorithms could outperform classical survival analysis methods. Methods: In this benchmarking study, two datasets were used to develop and compare different prognostic models for overall survival in pan-cancer populations: a nationwide EHR-derived de-identified database for training and in-sample testing and the OAK (phase III clinical trial) dataset for out-of-sample testing. A real-world database comprised 136K first-line treated cancer patients across multiple cancer types and was split into a 90% training and 10% testing dataset, respectively. The OAK dataset comprised 1,187 patients diagnosed with non-small cell lung cancer. To assess the effect of the covariate number on prognostic performance, we formed three feature sets with 27, 44 and 88 covariates. In terms of methods, we benchmarked ROPRO, a prognostic score based on the Cox model, against eight complex machine-learning models: regularized Cox, Random Survival Forests (RSF), Gradient Boosting (GB), DeepSurv (DS), Autoencoder (AE) and Super Learner (SL). The C-index was used as the performance metric to compare different models. Results: For in-sample testing on the real-world database the resulting C-index [95% CI] values for RSF 0.720 [0.716, 0.725], GB 0.722 [0.718, 0.727], DS 0.721 [0.717, 0.726] and lastly, SL 0.723 [0.718, 0.728] showed significantly better performance as compared to ROPRO 0.701 [0.696, 0.706]. Similar results were derived across all feature sets. However, for the out-of-sample validation on OAK, the stronger performance of the more complex models was not apparent anymore. Consistently, the increase in the number of prognostic covariates did not lead to an increase in model performance. Discussion: The stronger performance of the more complex models did not generalize when applied to an out-of-sample dataset. We hypothesize that future research may benefit by adding multimodal data to exploit advantages of more complex models. |
format | Online Article Text |
id | pubmed-8086599 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2021 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-80865992021-05-01 Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study Loureiro, Hugo Becker, Tim Bauer-Mehren, Anna Ahmidi, Narges Weberpals, Janick Front Artif Intell Artificial Intelligence Introduction: Prognostic scores are important tools in oncology to facilitate clinical decision-making based on patient characteristics. To date, classic survival analysis using Cox proportional hazards regression has been employed in the development of these prognostic scores. With the advance of analytical models, this study aimed to determine if more complex machine-learning algorithms could outperform classical survival analysis methods. Methods: In this benchmarking study, two datasets were used to develop and compare different prognostic models for overall survival in pan-cancer populations: a nationwide EHR-derived de-identified database for training and in-sample testing and the OAK (phase III clinical trial) dataset for out-of-sample testing. A real-world database comprised 136K first-line treated cancer patients across multiple cancer types and was split into a 90% training and 10% testing dataset, respectively. The OAK dataset comprised 1,187 patients diagnosed with non-small cell lung cancer. To assess the effect of the covariate number on prognostic performance, we formed three feature sets with 27, 44 and 88 covariates. In terms of methods, we benchmarked ROPRO, a prognostic score based on the Cox model, against eight complex machine-learning models: regularized Cox, Random Survival Forests (RSF), Gradient Boosting (GB), DeepSurv (DS), Autoencoder (AE) and Super Learner (SL). The C-index was used as the performance metric to compare different models. Results: For in-sample testing on the real-world database the resulting C-index [95% CI] values for RSF 0.720 [0.716, 0.725], GB 0.722 [0.718, 0.727], DS 0.721 [0.717, 0.726] and lastly, SL 0.723 [0.718, 0.728] showed significantly better performance as compared to ROPRO 0.701 [0.696, 0.706]. Similar results were derived across all feature sets. However, for the out-of-sample validation on OAK, the stronger performance of the more complex models was not apparent anymore. Consistently, the increase in the number of prognostic covariates did not lead to an increase in model performance. Discussion: The stronger performance of the more complex models did not generalize when applied to an out-of-sample dataset. We hypothesize that future research may benefit by adding multimodal data to exploit advantages of more complex models. Frontiers Media S.A. 2021-04-16 /pmc/articles/PMC8086599/ /pubmed/33937744 http://dx.doi.org/10.3389/frai.2021.625573 Text en Copyright © 2021 Loureiro, Becker, Bauer-Mehren, Ahmidi and Weberpals. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Artificial Intelligence Loureiro, Hugo Becker, Tim Bauer-Mehren, Anna Ahmidi, Narges Weberpals, Janick Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study |
title | Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study |
title_full | Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study |
title_fullStr | Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study |
title_full_unstemmed | Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study |
title_short | Artificial Intelligence for Prognostic Scores in Oncology: a Benchmarking Study |
title_sort | artificial intelligence for prognostic scores in oncology: a benchmarking study |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8086599/ https://www.ncbi.nlm.nih.gov/pubmed/33937744 http://dx.doi.org/10.3389/frai.2021.625573 |
work_keys_str_mv | AT loureirohugo artificialintelligenceforprognosticscoresinoncologyabenchmarkingstudy AT beckertim artificialintelligenceforprognosticscoresinoncologyabenchmarkingstudy AT bauermehrenanna artificialintelligenceforprognosticscoresinoncologyabenchmarkingstudy AT ahmidinarges artificialintelligenceforprognosticscoresinoncologyabenchmarkingstudy AT weberpalsjanick artificialintelligenceforprognosticscoresinoncologyabenchmarkingstudy |