Cargando…
Evaluating Modeling and Validation Strategies for Tooth Loss
Prediction models learn patterns from available data (training) and are then validated on new data (testing). Prediction modeling is increasingly common in dental research. We aimed to evaluate how different model development and validation steps affect the predictive performance of tooth loss predi...
Autores principales: | , , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
SAGE Publications
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6710618/ https://www.ncbi.nlm.nih.gov/pubmed/31361174 http://dx.doi.org/10.1177/0022034519864889 |
_version_ | 1783446376762310656 |
---|---|
author | Krois, J. Graetz, C. Holtfreter, B. Brinkmann, P. Kocher, T. Schwendicke, F. |
author_facet | Krois, J. Graetz, C. Holtfreter, B. Brinkmann, P. Kocher, T. Schwendicke, F. |
author_sort | Krois, J. |
collection | PubMed |
description | Prediction models learn patterns from available data (training) and are then validated on new data (testing). Prediction modeling is increasingly common in dental research. We aimed to evaluate how different model development and validation steps affect the predictive performance of tooth loss prediction models of patients with periodontitis. Two independent cohorts (627 patients, 11,651 teeth) were followed over a mean ± SD 18.2 ± 5.6 y (Kiel cohort) and 6.6 ± 2.9 y (Greifswald cohort). Tooth loss and 10 patient- and tooth-level predictors were recorded. The impact of different model development and validation steps was evaluated: 1) model complexity (logistic regression, recursive partitioning, random forest, extreme gradient boosting), 2) sample size (full data set or 10%, 25%, or 75% of cases dropped at random), 3) prediction periods (maximum 10, 15, or 20 y or uncensored), and 4) validation schemes (internal or external by centers/time). Tooth loss was generally a rare event (880 teeth were lost). All models showed limited sensitivity but high specificity. Patients’ age and tooth loss at baseline as well as probing pocket depths showed high variable importance. More complex models (random forest, extreme gradient boosting) had no consistent advantages over simpler ones (logistic regression, recursive partitioning). Internal validation (in sample) overestimated the predictive power (area under the curve up to 0.90), while external validation (out of sample) found lower areas under the curve (range 0.62 to 0.82). Reducing the sample size decreased the predictive power, particularly for more complex models. Censoring the prediction period had only limited impact. When the model was trained in one period and tested in another, model outcomes were similar to the base case, indicating temporal validation as a valid option. No model showed higher accuracy than the no-information rate. In conclusion, none of the developed models would be useful in a clinical setting, despite high accuracy. During modeling, rigorous development and external validation should be applied and reported accordingly. |
format | Online Article Text |
id | pubmed-6710618 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | SAGE Publications |
record_format | MEDLINE/PubMed |
spelling | pubmed-67106182019-09-17 Evaluating Modeling and Validation Strategies for Tooth Loss Krois, J. Graetz, C. Holtfreter, B. Brinkmann, P. Kocher, T. Schwendicke, F. J Dent Res Research Reports Prediction models learn patterns from available data (training) and are then validated on new data (testing). Prediction modeling is increasingly common in dental research. We aimed to evaluate how different model development and validation steps affect the predictive performance of tooth loss prediction models of patients with periodontitis. Two independent cohorts (627 patients, 11,651 teeth) were followed over a mean ± SD 18.2 ± 5.6 y (Kiel cohort) and 6.6 ± 2.9 y (Greifswald cohort). Tooth loss and 10 patient- and tooth-level predictors were recorded. The impact of different model development and validation steps was evaluated: 1) model complexity (logistic regression, recursive partitioning, random forest, extreme gradient boosting), 2) sample size (full data set or 10%, 25%, or 75% of cases dropped at random), 3) prediction periods (maximum 10, 15, or 20 y or uncensored), and 4) validation schemes (internal or external by centers/time). Tooth loss was generally a rare event (880 teeth were lost). All models showed limited sensitivity but high specificity. Patients’ age and tooth loss at baseline as well as probing pocket depths showed high variable importance. More complex models (random forest, extreme gradient boosting) had no consistent advantages over simpler ones (logistic regression, recursive partitioning). Internal validation (in sample) overestimated the predictive power (area under the curve up to 0.90), while external validation (out of sample) found lower areas under the curve (range 0.62 to 0.82). Reducing the sample size decreased the predictive power, particularly for more complex models. Censoring the prediction period had only limited impact. When the model was trained in one period and tested in another, model outcomes were similar to the base case, indicating temporal validation as a valid option. No model showed higher accuracy than the no-information rate. In conclusion, none of the developed models would be useful in a clinical setting, despite high accuracy. During modeling, rigorous development and external validation should be applied and reported accordingly. SAGE Publications 2019-07-30 2019-09 /pmc/articles/PMC6710618/ /pubmed/31361174 http://dx.doi.org/10.1177/0022034519864889 Text en © International & American Associations for Dental Research 2019 http://creativecommons.org/licenses/by/4.0/ This article is distributed under the terms of the Creative Commons Attribution 4.0 License (http://www.creativecommons.org/licenses/by/4.0/) which permits any use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/open-access-at-sage). |
spellingShingle | Research Reports Krois, J. Graetz, C. Holtfreter, B. Brinkmann, P. Kocher, T. Schwendicke, F. Evaluating Modeling and Validation Strategies for Tooth Loss |
title | Evaluating Modeling and Validation Strategies for Tooth Loss |
title_full | Evaluating Modeling and Validation Strategies for Tooth Loss |
title_fullStr | Evaluating Modeling and Validation Strategies for Tooth Loss |
title_full_unstemmed | Evaluating Modeling and Validation Strategies for Tooth Loss |
title_short | Evaluating Modeling and Validation Strategies for Tooth Loss |
title_sort | evaluating modeling and validation strategies for tooth loss |
topic | Research Reports |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6710618/ https://www.ncbi.nlm.nih.gov/pubmed/31361174 http://dx.doi.org/10.1177/0022034519864889 |
work_keys_str_mv | AT kroisj evaluatingmodelingandvalidationstrategiesfortoothloss AT graetzc evaluatingmodelingandvalidationstrategiesfortoothloss AT holtfreterb evaluatingmodelingandvalidationstrategiesfortoothloss AT brinkmannp evaluatingmodelingandvalidationstrategiesfortoothloss AT kochert evaluatingmodelingandvalidationstrategiesfortoothloss AT schwendickef evaluatingmodelingandvalidationstrategiesfortoothloss |