Cargando…
Differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia - lessons for translation into clinical practice
Artificial intelligence (AI) approaches pose a great opportunity for individualized, pre-symptomatic disease diagnosis which plays a key role in the context of personalized, predictive, and finally preventive medicine (PPPM). However, to translate PPPM into clinical practice, it is of utmost importa...
Autores principales: | , , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer International Publishing
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7429672/ https://www.ncbi.nlm.nih.gov/pubmed/32843907 http://dx.doi.org/10.1007/s13167-020-00216-z |
_version_ | 1783571294300667904 |
---|---|
author | Birkenbihl, Colin Emon, Mohammad Asif Vrooman, Henri Westwood, Sarah Lovestone, Simon Hofmann-Apitius, Martin Fröhlich, Holger |
author_facet | Birkenbihl, Colin Emon, Mohammad Asif Vrooman, Henri Westwood, Sarah Lovestone, Simon Hofmann-Apitius, Martin Fröhlich, Holger |
author_sort | Birkenbihl, Colin |
collection | PubMed |
description | Artificial intelligence (AI) approaches pose a great opportunity for individualized, pre-symptomatic disease diagnosis which plays a key role in the context of personalized, predictive, and finally preventive medicine (PPPM). However, to translate PPPM into clinical practice, it is of utmost importance that AI-based models are carefully validated. The validation process comprises several steps, one of which is testing the model on patient-level data from an independent clinical cohort study. However, recruitment criteria can bias statistical analysis of cohort study data and impede model application beyond the training data. To evaluate whether and how data from independent clinical cohort studies differ from each other, this study systematically compares the datasets collected from two major dementia cohorts, namely, the Alzheimer’s Disease Neuroimaging Initiative (ADNI) and AddNeuroMed. The presented comparison was conducted on individual feature level and revealed significant differences among both cohorts. Such systematic deviations can potentially hamper the generalizability of results which were based on a single cohort dataset. Despite identified differences, validation of a previously published, ADNI trained model for prediction of personalized dementia risk scores on 244 AddNeuroMed subjects was successful: External validation resulted in a high prediction performance of above 80% area under receiver operator characteristic curve up to 6 years before dementia diagnosis. Propensity score matching identified a subset of patients from AddNeuroMed, which showed significantly smaller demographic differences to ADNI. For these patients, an even higher prediction performance was achieved, which demonstrates the influence systematic differences between cohorts can have on validation results. In conclusion, this study exposes challenges in external validation of AI models on cohort study data and is one of the rare cases in the neurology field in which such external validation was performed. The presented model represents a proof of concept that reliable models for personalized predictive diagnostics are feasible, which, in turn, could lead to adequate disease prevention and hereby enable the PPPM paradigm in the dementia field. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (10.1007/s13167-020-00216-z) contains supplementary material, which is available to authorized users. |
format | Online Article Text |
id | pubmed-7429672 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
publisher | Springer International Publishing |
record_format | MEDLINE/PubMed |
spelling | pubmed-74296722020-08-24 Differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia - lessons for translation into clinical practice Birkenbihl, Colin Emon, Mohammad Asif Vrooman, Henri Westwood, Sarah Lovestone, Simon Hofmann-Apitius, Martin Fröhlich, Holger EPMA J Research Artificial intelligence (AI) approaches pose a great opportunity for individualized, pre-symptomatic disease diagnosis which plays a key role in the context of personalized, predictive, and finally preventive medicine (PPPM). However, to translate PPPM into clinical practice, it is of utmost importance that AI-based models are carefully validated. The validation process comprises several steps, one of which is testing the model on patient-level data from an independent clinical cohort study. However, recruitment criteria can bias statistical analysis of cohort study data and impede model application beyond the training data. To evaluate whether and how data from independent clinical cohort studies differ from each other, this study systematically compares the datasets collected from two major dementia cohorts, namely, the Alzheimer’s Disease Neuroimaging Initiative (ADNI) and AddNeuroMed. The presented comparison was conducted on individual feature level and revealed significant differences among both cohorts. Such systematic deviations can potentially hamper the generalizability of results which were based on a single cohort dataset. Despite identified differences, validation of a previously published, ADNI trained model for prediction of personalized dementia risk scores on 244 AddNeuroMed subjects was successful: External validation resulted in a high prediction performance of above 80% area under receiver operator characteristic curve up to 6 years before dementia diagnosis. Propensity score matching identified a subset of patients from AddNeuroMed, which showed significantly smaller demographic differences to ADNI. For these patients, an even higher prediction performance was achieved, which demonstrates the influence systematic differences between cohorts can have on validation results. In conclusion, this study exposes challenges in external validation of AI models on cohort study data and is one of the rare cases in the neurology field in which such external validation was performed. The presented model represents a proof of concept that reliable models for personalized predictive diagnostics are feasible, which, in turn, could lead to adequate disease prevention and hereby enable the PPPM paradigm in the dementia field. ELECTRONIC SUPPLEMENTARY MATERIAL: The online version of this article (10.1007/s13167-020-00216-z) contains supplementary material, which is available to authorized users. Springer International Publishing 2020-06-22 /pmc/articles/PMC7429672/ /pubmed/32843907 http://dx.doi.org/10.1007/s13167-020-00216-z Text en © The Author(s) 2020 Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. |
spellingShingle | Research Birkenbihl, Colin Emon, Mohammad Asif Vrooman, Henri Westwood, Sarah Lovestone, Simon Hofmann-Apitius, Martin Fröhlich, Holger Differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia - lessons for translation into clinical practice |
title | Differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia - lessons for translation into clinical practice |
title_full | Differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia - lessons for translation into clinical practice |
title_fullStr | Differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia - lessons for translation into clinical practice |
title_full_unstemmed | Differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia - lessons for translation into clinical practice |
title_short | Differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia - lessons for translation into clinical practice |
title_sort | differences in cohort study data affect external validation of artificial intelligence models for predictive diagnostics of dementia - lessons for translation into clinical practice |
topic | Research |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7429672/ https://www.ncbi.nlm.nih.gov/pubmed/32843907 http://dx.doi.org/10.1007/s13167-020-00216-z |
work_keys_str_mv | AT birkenbihlcolin differencesincohortstudydataaffectexternalvalidationofartificialintelligencemodelsforpredictivediagnosticsofdementialessonsfortranslationintoclinicalpractice AT emonmohammadasif differencesincohortstudydataaffectexternalvalidationofartificialintelligencemodelsforpredictivediagnosticsofdementialessonsfortranslationintoclinicalpractice AT vroomanhenri differencesincohortstudydataaffectexternalvalidationofartificialintelligencemodelsforpredictivediagnosticsofdementialessonsfortranslationintoclinicalpractice AT westwoodsarah differencesincohortstudydataaffectexternalvalidationofartificialintelligencemodelsforpredictivediagnosticsofdementialessonsfortranslationintoclinicalpractice AT lovestonesimon differencesincohortstudydataaffectexternalvalidationofartificialintelligencemodelsforpredictivediagnosticsofdementialessonsfortranslationintoclinicalpractice AT differencesincohortstudydataaffectexternalvalidationofartificialintelligencemodelsforpredictivediagnosticsofdementialessonsfortranslationintoclinicalpractice AT hofmannapitiusmartin differencesincohortstudydataaffectexternalvalidationofartificialintelligencemodelsforpredictivediagnosticsofdementialessonsfortranslationintoclinicalpractice AT frohlichholger differencesincohortstudydataaffectexternalvalidationofartificialintelligencemodelsforpredictivediagnosticsofdementialessonsfortranslationintoclinicalpractice AT differencesincohortstudydataaffectexternalvalidationofartificialintelligencemodelsforpredictivediagnosticsofdementialessonsfortranslationintoclinicalpractice |