Cargando…

Calibration of medical diagnostic classifier scores to the probability of disease

Scores produced by statistical classifiers in many clinical decision support systems and other medical diagnostic devices are generally on an arbitrary scale, so the clinical meaning of these scores is unclear. Calibration of classifier scores to a meaningful scale such as the probability of disease...

Descripción completa

Detalles Bibliográficos
Autores principales: Chen, Weijie, Sahiner, Berkman, Samuelson, Frank, Pezeshk, Aria, Petrick, Nicholas
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2016
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5548655/
https://www.ncbi.nlm.nih.gov/pubmed/27507287
http://dx.doi.org/10.1177/0962280216661371
_version_ 1783255848803368960
author Chen, Weijie
Sahiner, Berkman
Samuelson, Frank
Pezeshk, Aria
Petrick, Nicholas
author_facet Chen, Weijie
Sahiner, Berkman
Samuelson, Frank
Pezeshk, Aria
Petrick, Nicholas
author_sort Chen, Weijie
collection PubMed
description Scores produced by statistical classifiers in many clinical decision support systems and other medical diagnostic devices are generally on an arbitrary scale, so the clinical meaning of these scores is unclear. Calibration of classifier scores to a meaningful scale such as the probability of disease is potentially useful when such scores are used by a physician. In this work, we investigated three methods (parametric, semi-parametric, and non-parametric) for calibrating classifier scores to the probability of disease scale and developed uncertainty estimation techniques for these methods. We showed that classifier scores on arbitrary scales can be calibrated to the probability of disease scale without affecting their discrimination performance. With a finite dataset to train the calibration function, it is important to accompany the probability estimate with its confidence interval. Our simulations indicate that, when a dataset used for finding the transformation for calibration is also used for estimating the performance of calibration, the resubstitution bias exists for a performance metric involving the truth states in evaluating the calibration performance. However, the bias is small for the parametric and semi-parametric methods when the sample size is moderate to large (>100 per class).
format Online
Article
Text
id pubmed-5548655
institution National Center for Biotechnology Information
language English
publishDate 2016
record_format MEDLINE/PubMed
spelling pubmed-55486552018-05-01 Calibration of medical diagnostic classifier scores to the probability of disease Chen, Weijie Sahiner, Berkman Samuelson, Frank Pezeshk, Aria Petrick, Nicholas Stat Methods Med Res Article Scores produced by statistical classifiers in many clinical decision support systems and other medical diagnostic devices are generally on an arbitrary scale, so the clinical meaning of these scores is unclear. Calibration of classifier scores to a meaningful scale such as the probability of disease is potentially useful when such scores are used by a physician. In this work, we investigated three methods (parametric, semi-parametric, and non-parametric) for calibrating classifier scores to the probability of disease scale and developed uncertainty estimation techniques for these methods. We showed that classifier scores on arbitrary scales can be calibrated to the probability of disease scale without affecting their discrimination performance. With a finite dataset to train the calibration function, it is important to accompany the probability estimate with its confidence interval. Our simulations indicate that, when a dataset used for finding the transformation for calibration is also used for estimating the performance of calibration, the resubstitution bias exists for a performance metric involving the truth states in evaluating the calibration performance. However, the bias is small for the parametric and semi-parametric methods when the sample size is moderate to large (>100 per class). 2016-08-08 2018-05 /pmc/articles/PMC5548655/ /pubmed/27507287 http://dx.doi.org/10.1177/0962280216661371 Text en http://creativecommons.org/licenses/by/2.0/ Reprints and permissions: sagepub.co.uk/journalsPermissions.nav (http://sagepub.co.uk/journalsPermissions.nav)
spellingShingle Article
Chen, Weijie
Sahiner, Berkman
Samuelson, Frank
Pezeshk, Aria
Petrick, Nicholas
Calibration of medical diagnostic classifier scores to the probability of disease
title Calibration of medical diagnostic classifier scores to the probability of disease
title_full Calibration of medical diagnostic classifier scores to the probability of disease
title_fullStr Calibration of medical diagnostic classifier scores to the probability of disease
title_full_unstemmed Calibration of medical diagnostic classifier scores to the probability of disease
title_short Calibration of medical diagnostic classifier scores to the probability of disease
title_sort calibration of medical diagnostic classifier scores to the probability of disease
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5548655/
https://www.ncbi.nlm.nih.gov/pubmed/27507287
http://dx.doi.org/10.1177/0962280216661371
work_keys_str_mv AT chenweijie calibrationofmedicaldiagnosticclassifierscorestotheprobabilityofdisease
AT sahinerberkman calibrationofmedicaldiagnosticclassifierscorestotheprobabilityofdisease
AT samuelsonfrank calibrationofmedicaldiagnosticclassifierscorestotheprobabilityofdisease
AT pezeshkaria calibrationofmedicaldiagnosticclassifierscorestotheprobabilityofdisease
AT petricknicholas calibrationofmedicaldiagnosticclassifierscorestotheprobabilityofdisease