Cargando…

Improving reproducibility by using high-throughput observational studies with empirical calibration

Concerns over reproducibility in science extend to research using existing healthcare data; many observational studies investigating the same topic produce conflicting results, even when using the same data. To address this problem, we propose a paradigm shift. The current paradigm centres on genera...

Descripción completa

Detalles Bibliográficos
Autores principales: Schuemie, Martijn J., Ryan, Patrick B., Hripcsak, George, Madigan, David, Suchard, Marc A.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: The Royal Society Publishing 2018
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6107542/
https://www.ncbi.nlm.nih.gov/pubmed/30082302
http://dx.doi.org/10.1098/rsta.2017.0356
_version_ 1783349983406194688
author Schuemie, Martijn J.
Ryan, Patrick B.
Hripcsak, George
Madigan, David
Suchard, Marc A.
author_facet Schuemie, Martijn J.
Ryan, Patrick B.
Hripcsak, George
Madigan, David
Suchard, Marc A.
author_sort Schuemie, Martijn J.
collection PubMed
description Concerns over reproducibility in science extend to research using existing healthcare data; many observational studies investigating the same topic produce conflicting results, even when using the same data. To address this problem, we propose a paradigm shift. The current paradigm centres on generating one estimate at a time using a unique study design with unknown reliability and publishing (or not) one estimate at a time. The new paradigm advocates for high-throughput observational studies using consistent and standardized methods, allowing evaluation, calibration and unbiased dissemination to generate a more reliable and complete evidence base. We demonstrate this new paradigm by comparing all depression treatments for a set of outcomes, producing 17 718 hazard ratios, each using methodology on par with current best practice. We furthermore include control hypotheses to evaluate and calibrate our evidence generation process. Results show good transitivity and consistency between databases, and agree with four out of the five findings from clinical trials. The distribution of effect size estimates reported in the literature reveals an absence of small or null effects, with a sharp cut-off at p = 0.05. No such phenomena were observed in our results, suggesting more complete and more reliable evidence. This article is part of a discussion meeting issue ‘The growing ubiquity of algorithms in society: implications, impacts and innovations’.
format Online
Article
Text
id pubmed-6107542
institution National Center for Biotechnology Information
language English
publishDate 2018
publisher The Royal Society Publishing
record_format MEDLINE/PubMed
spelling pubmed-61075422018-08-24 Improving reproducibility by using high-throughput observational studies with empirical calibration Schuemie, Martijn J. Ryan, Patrick B. Hripcsak, George Madigan, David Suchard, Marc A. Philos Trans A Math Phys Eng Sci Articles Concerns over reproducibility in science extend to research using existing healthcare data; many observational studies investigating the same topic produce conflicting results, even when using the same data. To address this problem, we propose a paradigm shift. The current paradigm centres on generating one estimate at a time using a unique study design with unknown reliability and publishing (or not) one estimate at a time. The new paradigm advocates for high-throughput observational studies using consistent and standardized methods, allowing evaluation, calibration and unbiased dissemination to generate a more reliable and complete evidence base. We demonstrate this new paradigm by comparing all depression treatments for a set of outcomes, producing 17 718 hazard ratios, each using methodology on par with current best practice. We furthermore include control hypotheses to evaluate and calibrate our evidence generation process. Results show good transitivity and consistency between databases, and agree with four out of the five findings from clinical trials. The distribution of effect size estimates reported in the literature reveals an absence of small or null effects, with a sharp cut-off at p = 0.05. No such phenomena were observed in our results, suggesting more complete and more reliable evidence. This article is part of a discussion meeting issue ‘The growing ubiquity of algorithms in society: implications, impacts and innovations’. The Royal Society Publishing 2018-09-13 2018-08-06 /pmc/articles/PMC6107542/ /pubmed/30082302 http://dx.doi.org/10.1098/rsta.2017.0356 Text en © 2018 The Authors. http://creativecommons.org/licenses/by/4.0/ Published by the Royal Society under the terms of the Creative Commons Attribution License http://creativecommons.org/licenses/by/4.0/, which permits unrestricted use, provided the original author and source are credited.
spellingShingle Articles
Schuemie, Martijn J.
Ryan, Patrick B.
Hripcsak, George
Madigan, David
Suchard, Marc A.
Improving reproducibility by using high-throughput observational studies with empirical calibration
title Improving reproducibility by using high-throughput observational studies with empirical calibration
title_full Improving reproducibility by using high-throughput observational studies with empirical calibration
title_fullStr Improving reproducibility by using high-throughput observational studies with empirical calibration
title_full_unstemmed Improving reproducibility by using high-throughput observational studies with empirical calibration
title_short Improving reproducibility by using high-throughput observational studies with empirical calibration
title_sort improving reproducibility by using high-throughput observational studies with empirical calibration
topic Articles
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6107542/
https://www.ncbi.nlm.nih.gov/pubmed/30082302
http://dx.doi.org/10.1098/rsta.2017.0356
work_keys_str_mv AT schuemiemartijnj improvingreproducibilitybyusinghighthroughputobservationalstudieswithempiricalcalibration
AT ryanpatrickb improvingreproducibilitybyusinghighthroughputobservationalstudieswithempiricalcalibration
AT hripcsakgeorge improvingreproducibilitybyusinghighthroughputobservationalstudieswithempiricalcalibration
AT madigandavid improvingreproducibilitybyusinghighthroughputobservationalstudieswithempiricalcalibration
AT suchardmarca improvingreproducibilitybyusinghighthroughputobservationalstudieswithempiricalcalibration