Cargando…

Evaluating Modeling and Validation Strategies for Tooth Loss

Prediction models learn patterns from available data (training) and are then validated on new data (testing). Prediction modeling is increasingly common in dental research. We aimed to evaluate how different model development and validation steps affect the predictive performance of tooth loss predi...

Descripción completa

Detalles Bibliográficos
Autores principales: Krois, J., Graetz, C., Holtfreter, B., Brinkmann, P., Kocher, T., Schwendicke, F.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: SAGE Publications 2019
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6710618/
https://www.ncbi.nlm.nih.gov/pubmed/31361174
http://dx.doi.org/10.1177/0022034519864889
_version_ 1783446376762310656
author Krois, J.
Graetz, C.
Holtfreter, B.
Brinkmann, P.
Kocher, T.
Schwendicke, F.
author_facet Krois, J.
Graetz, C.
Holtfreter, B.
Brinkmann, P.
Kocher, T.
Schwendicke, F.
author_sort Krois, J.
collection PubMed
description Prediction models learn patterns from available data (training) and are then validated on new data (testing). Prediction modeling is increasingly common in dental research. We aimed to evaluate how different model development and validation steps affect the predictive performance of tooth loss prediction models of patients with periodontitis. Two independent cohorts (627 patients, 11,651 teeth) were followed over a mean ± SD 18.2 ± 5.6 y (Kiel cohort) and 6.6 ± 2.9 y (Greifswald cohort). Tooth loss and 10 patient- and tooth-level predictors were recorded. The impact of different model development and validation steps was evaluated: 1) model complexity (logistic regression, recursive partitioning, random forest, extreme gradient boosting), 2) sample size (full data set or 10%, 25%, or 75% of cases dropped at random), 3) prediction periods (maximum 10, 15, or 20 y or uncensored), and 4) validation schemes (internal or external by centers/time). Tooth loss was generally a rare event (880 teeth were lost). All models showed limited sensitivity but high specificity. Patients’ age and tooth loss at baseline as well as probing pocket depths showed high variable importance. More complex models (random forest, extreme gradient boosting) had no consistent advantages over simpler ones (logistic regression, recursive partitioning). Internal validation (in sample) overestimated the predictive power (area under the curve up to 0.90), while external validation (out of sample) found lower areas under the curve (range 0.62 to 0.82). Reducing the sample size decreased the predictive power, particularly for more complex models. Censoring the prediction period had only limited impact. When the model was trained in one period and tested in another, model outcomes were similar to the base case, indicating temporal validation as a valid option. No model showed higher accuracy than the no-information rate. In conclusion, none of the developed models would be useful in a clinical setting, despite high accuracy. During modeling, rigorous development and external validation should be applied and reported accordingly.
format Online
Article
Text
id pubmed-6710618
institution National Center for Biotechnology Information
language English
publishDate 2019
publisher SAGE Publications
record_format MEDLINE/PubMed
spelling pubmed-67106182019-09-17 Evaluating Modeling and Validation Strategies for Tooth Loss Krois, J. Graetz, C. Holtfreter, B. Brinkmann, P. Kocher, T. Schwendicke, F. J Dent Res Research Reports Prediction models learn patterns from available data (training) and are then validated on new data (testing). Prediction modeling is increasingly common in dental research. We aimed to evaluate how different model development and validation steps affect the predictive performance of tooth loss prediction models of patients with periodontitis. Two independent cohorts (627 patients, 11,651 teeth) were followed over a mean ± SD 18.2 ± 5.6 y (Kiel cohort) and 6.6 ± 2.9 y (Greifswald cohort). Tooth loss and 10 patient- and tooth-level predictors were recorded. The impact of different model development and validation steps was evaluated: 1) model complexity (logistic regression, recursive partitioning, random forest, extreme gradient boosting), 2) sample size (full data set or 10%, 25%, or 75% of cases dropped at random), 3) prediction periods (maximum 10, 15, or 20 y or uncensored), and 4) validation schemes (internal or external by centers/time). Tooth loss was generally a rare event (880 teeth were lost). All models showed limited sensitivity but high specificity. Patients’ age and tooth loss at baseline as well as probing pocket depths showed high variable importance. More complex models (random forest, extreme gradient boosting) had no consistent advantages over simpler ones (logistic regression, recursive partitioning). Internal validation (in sample) overestimated the predictive power (area under the curve up to 0.90), while external validation (out of sample) found lower areas under the curve (range 0.62 to 0.82). Reducing the sample size decreased the predictive power, particularly for more complex models. Censoring the prediction period had only limited impact. When the model was trained in one period and tested in another, model outcomes were similar to the base case, indicating temporal validation as a valid option. No model showed higher accuracy than the no-information rate. In conclusion, none of the developed models would be useful in a clinical setting, despite high accuracy. During modeling, rigorous development and external validation should be applied and reported accordingly. SAGE Publications 2019-07-30 2019-09 /pmc/articles/PMC6710618/ /pubmed/31361174 http://dx.doi.org/10.1177/0022034519864889 Text en © International & American Associations for Dental Research 2019 http://creativecommons.org/licenses/by/4.0/ This article is distributed under the terms of the Creative Commons Attribution 4.0 License (http://www.creativecommons.org/licenses/by/4.0/) which permits any use, reproduction and distribution of the work without further permission provided the original work is attributed as specified on the SAGE and Open Access pages (https://us.sagepub.com/en-us/nam/open-access-at-sage).
spellingShingle Research Reports
Krois, J.
Graetz, C.
Holtfreter, B.
Brinkmann, P.
Kocher, T.
Schwendicke, F.
Evaluating Modeling and Validation Strategies for Tooth Loss
title Evaluating Modeling and Validation Strategies for Tooth Loss
title_full Evaluating Modeling and Validation Strategies for Tooth Loss
title_fullStr Evaluating Modeling and Validation Strategies for Tooth Loss
title_full_unstemmed Evaluating Modeling and Validation Strategies for Tooth Loss
title_short Evaluating Modeling and Validation Strategies for Tooth Loss
title_sort evaluating modeling and validation strategies for tooth loss
topic Research Reports
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6710618/
https://www.ncbi.nlm.nih.gov/pubmed/31361174
http://dx.doi.org/10.1177/0022034519864889
work_keys_str_mv AT kroisj evaluatingmodelingandvalidationstrategiesfortoothloss
AT graetzc evaluatingmodelingandvalidationstrategiesfortoothloss
AT holtfreterb evaluatingmodelingandvalidationstrategiesfortoothloss
AT brinkmannp evaluatingmodelingandvalidationstrategiesfortoothloss
AT kochert evaluatingmodelingandvalidationstrategiesfortoothloss
AT schwendickef evaluatingmodelingandvalidationstrategiesfortoothloss