Cargando…
Calibration of medical diagnostic classifier scores to the probability of disease
Scores produced by statistical classifiers in many clinical decision support systems and other medical diagnostic devices are generally on an arbitrary scale, so the clinical meaning of these scores is unclear. Calibration of classifier scores to a meaningful scale such as the probability of disease...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2016
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5548655/ https://www.ncbi.nlm.nih.gov/pubmed/27507287 http://dx.doi.org/10.1177/0962280216661371 |
_version_ | 1783255848803368960 |
---|---|
author | Chen, Weijie Sahiner, Berkman Samuelson, Frank Pezeshk, Aria Petrick, Nicholas |
author_facet | Chen, Weijie Sahiner, Berkman Samuelson, Frank Pezeshk, Aria Petrick, Nicholas |
author_sort | Chen, Weijie |
collection | PubMed |
description | Scores produced by statistical classifiers in many clinical decision support systems and other medical diagnostic devices are generally on an arbitrary scale, so the clinical meaning of these scores is unclear. Calibration of classifier scores to a meaningful scale such as the probability of disease is potentially useful when such scores are used by a physician. In this work, we investigated three methods (parametric, semi-parametric, and non-parametric) for calibrating classifier scores to the probability of disease scale and developed uncertainty estimation techniques for these methods. We showed that classifier scores on arbitrary scales can be calibrated to the probability of disease scale without affecting their discrimination performance. With a finite dataset to train the calibration function, it is important to accompany the probability estimate with its confidence interval. Our simulations indicate that, when a dataset used for finding the transformation for calibration is also used for estimating the performance of calibration, the resubstitution bias exists for a performance metric involving the truth states in evaluating the calibration performance. However, the bias is small for the parametric and semi-parametric methods when the sample size is moderate to large (>100 per class). |
format | Online Article Text |
id | pubmed-5548655 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2016 |
record_format | MEDLINE/PubMed |
spelling | pubmed-55486552018-05-01 Calibration of medical diagnostic classifier scores to the probability of disease Chen, Weijie Sahiner, Berkman Samuelson, Frank Pezeshk, Aria Petrick, Nicholas Stat Methods Med Res Article Scores produced by statistical classifiers in many clinical decision support systems and other medical diagnostic devices are generally on an arbitrary scale, so the clinical meaning of these scores is unclear. Calibration of classifier scores to a meaningful scale such as the probability of disease is potentially useful when such scores are used by a physician. In this work, we investigated three methods (parametric, semi-parametric, and non-parametric) for calibrating classifier scores to the probability of disease scale and developed uncertainty estimation techniques for these methods. We showed that classifier scores on arbitrary scales can be calibrated to the probability of disease scale without affecting their discrimination performance. With a finite dataset to train the calibration function, it is important to accompany the probability estimate with its confidence interval. Our simulations indicate that, when a dataset used for finding the transformation for calibration is also used for estimating the performance of calibration, the resubstitution bias exists for a performance metric involving the truth states in evaluating the calibration performance. However, the bias is small for the parametric and semi-parametric methods when the sample size is moderate to large (>100 per class). 2016-08-08 2018-05 /pmc/articles/PMC5548655/ /pubmed/27507287 http://dx.doi.org/10.1177/0962280216661371 Text en http://creativecommons.org/licenses/by/2.0/ Reprints and permissions: sagepub.co.uk/journalsPermissions.nav (http://sagepub.co.uk/journalsPermissions.nav) |
spellingShingle | Article Chen, Weijie Sahiner, Berkman Samuelson, Frank Pezeshk, Aria Petrick, Nicholas Calibration of medical diagnostic classifier scores to the probability of disease |
title | Calibration of medical diagnostic classifier scores to the probability of disease |
title_full | Calibration of medical diagnostic classifier scores to the probability of disease |
title_fullStr | Calibration of medical diagnostic classifier scores to the probability of disease |
title_full_unstemmed | Calibration of medical diagnostic classifier scores to the probability of disease |
title_short | Calibration of medical diagnostic classifier scores to the probability of disease |
title_sort | calibration of medical diagnostic classifier scores to the probability of disease |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5548655/ https://www.ncbi.nlm.nih.gov/pubmed/27507287 http://dx.doi.org/10.1177/0962280216661371 |
work_keys_str_mv | AT chenweijie calibrationofmedicaldiagnosticclassifierscorestotheprobabilityofdisease AT sahinerberkman calibrationofmedicaldiagnosticclassifierscorestotheprobabilityofdisease AT samuelsonfrank calibrationofmedicaldiagnosticclassifierscorestotheprobabilityofdisease AT pezeshkaria calibrationofmedicaldiagnosticclassifierscorestotheprobabilityofdisease AT petricknicholas calibrationofmedicaldiagnosticclassifierscorestotheprobabilityofdisease |