Cargando…
The effect of uncertainty in patient classification on diagnostic performance estimations
BACKGROUND: The performance of a new diagnostic test is typically evaluated against a comparator which is assumed to correspond closely to some true state of interest. Judgments about the new test’s performance are based on the differences between the outputs of the test and comparator. It is common...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2019
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6530857/ https://www.ncbi.nlm.nih.gov/pubmed/31116772 http://dx.doi.org/10.1371/journal.pone.0217146 |
_version_ | 1783420712430600192 |
---|---|
author | McHugh, Leo C. Snyder, Kevin Yager, Thomas D. |
author_facet | McHugh, Leo C. Snyder, Kevin Yager, Thomas D. |
author_sort | McHugh, Leo C. |
collection | PubMed |
description | BACKGROUND: The performance of a new diagnostic test is typically evaluated against a comparator which is assumed to correspond closely to some true state of interest. Judgments about the new test’s performance are based on the differences between the outputs of the test and comparator. It is commonly assumed that a small amount of uncertainty in the comparator’s classifications will negligibly affect the measured performance of a diagnostic test. METHODS: Simulated datasets were generated to represent typical diagnostic scenarios. Comparator noise was introduced in the form of random misclassifications, and the effect on the apparent performance of the diagnostic test was determined. An actual dataset from a clinical trial on a new diagnostic test for sepsis was also analyzed. RESULTS: We demonstrate that as little as 5% misclassification of patients by the comparator can be enough to statistically invalidate performance estimates such as sensitivity, specificity and area under the receiver operating characteristic curve, if this uncertainty is not measured and taken into account. This distortion effect is found to increase non-linearly with comparator uncertainty, under some common diagnostic scenarios. For clinical populations exhibiting high degrees of classification uncertainty, failure to measure and account for this effect will introduce significant risks of drawing false conclusions. The effect of classification uncertainty is magnified further for high performing tests that would otherwise reach near-perfection in diagnostic evaluation trials. A requirement of very high diagnostic performance for clinical adoption, such as a 99% sensitivity, can be rendered nearly unachievable even for a perfect test, if the comparator diagnosis contains even small amounts of uncertainty. This paper and an accompanying online simulation tool demonstrate the effect of classification uncertainty on the apparent performance of tests across a range of typical diagnostic scenarios. Both simulated and real datasets are used to show the degradation of apparent test performance as comparator uncertainty increases. CONCLUSIONS: Overall, a 5% or greater misclassification rate by the comparator can lead to significant underestimation of true test performance. An online simulation tool allows researchers to explore this effect using their own trial parameters (https://imperfect-gold-standard.shinyapps.io/classification-noise/) and the source code is freely available (https://github.com/ksny/Imperfect-Gold-Standard). |
format | Online Article Text |
id | pubmed-6530857 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2019 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-65308572019-05-31 The effect of uncertainty in patient classification on diagnostic performance estimations McHugh, Leo C. Snyder, Kevin Yager, Thomas D. PLoS One Research Article BACKGROUND: The performance of a new diagnostic test is typically evaluated against a comparator which is assumed to correspond closely to some true state of interest. Judgments about the new test’s performance are based on the differences between the outputs of the test and comparator. It is commonly assumed that a small amount of uncertainty in the comparator’s classifications will negligibly affect the measured performance of a diagnostic test. METHODS: Simulated datasets were generated to represent typical diagnostic scenarios. Comparator noise was introduced in the form of random misclassifications, and the effect on the apparent performance of the diagnostic test was determined. An actual dataset from a clinical trial on a new diagnostic test for sepsis was also analyzed. RESULTS: We demonstrate that as little as 5% misclassification of patients by the comparator can be enough to statistically invalidate performance estimates such as sensitivity, specificity and area under the receiver operating characteristic curve, if this uncertainty is not measured and taken into account. This distortion effect is found to increase non-linearly with comparator uncertainty, under some common diagnostic scenarios. For clinical populations exhibiting high degrees of classification uncertainty, failure to measure and account for this effect will introduce significant risks of drawing false conclusions. The effect of classification uncertainty is magnified further for high performing tests that would otherwise reach near-perfection in diagnostic evaluation trials. A requirement of very high diagnostic performance for clinical adoption, such as a 99% sensitivity, can be rendered nearly unachievable even for a perfect test, if the comparator diagnosis contains even small amounts of uncertainty. This paper and an accompanying online simulation tool demonstrate the effect of classification uncertainty on the apparent performance of tests across a range of typical diagnostic scenarios. Both simulated and real datasets are used to show the degradation of apparent test performance as comparator uncertainty increases. CONCLUSIONS: Overall, a 5% or greater misclassification rate by the comparator can lead to significant underestimation of true test performance. An online simulation tool allows researchers to explore this effect using their own trial parameters (https://imperfect-gold-standard.shinyapps.io/classification-noise/) and the source code is freely available (https://github.com/ksny/Imperfect-Gold-Standard). Public Library of Science 2019-05-22 /pmc/articles/PMC6530857/ /pubmed/31116772 http://dx.doi.org/10.1371/journal.pone.0217146 Text en https://creativecommons.org/publicdomain/zero/1.0/ This is an open access article, free of all copyright, and may be freely reproduced, distributed, transmitted, modified, built upon, or otherwise used by anyone for any lawful purpose. The work is made available under the Creative Commons CC0 (https://creativecommons.org/publicdomain/zero/1.0/) public domain dedication. |
spellingShingle | Research Article McHugh, Leo C. Snyder, Kevin Yager, Thomas D. The effect of uncertainty in patient classification on diagnostic performance estimations |
title | The effect of uncertainty in patient classification on diagnostic performance estimations |
title_full | The effect of uncertainty in patient classification on diagnostic performance estimations |
title_fullStr | The effect of uncertainty in patient classification on diagnostic performance estimations |
title_full_unstemmed | The effect of uncertainty in patient classification on diagnostic performance estimations |
title_short | The effect of uncertainty in patient classification on diagnostic performance estimations |
title_sort | effect of uncertainty in patient classification on diagnostic performance estimations |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6530857/ https://www.ncbi.nlm.nih.gov/pubmed/31116772 http://dx.doi.org/10.1371/journal.pone.0217146 |
work_keys_str_mv | AT mchughleoc theeffectofuncertaintyinpatientclassificationondiagnosticperformanceestimations AT snyderkevin theeffectofuncertaintyinpatientclassificationondiagnosticperformanceestimations AT yagerthomasd theeffectofuncertaintyinpatientclassificationondiagnosticperformanceestimations AT mchughleoc effectofuncertaintyinpatientclassificationondiagnosticperformanceestimations AT snyderkevin effectofuncertaintyinpatientclassificationondiagnosticperformanceestimations AT yagerthomasd effectofuncertaintyinpatientclassificationondiagnosticperformanceestimations |