Cargando…
Improving reproducibility by using high-throughput observational studies with empirical calibration
Concerns over reproducibility in science extend to research using existing healthcare data; many observational studies investigating the same topic produce conflicting results, even when using the same data. To address this problem, we propose a paradigm shift. The current paradigm centres on genera...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
The Royal Society Publishing
2018
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6107542/ https://www.ncbi.nlm.nih.gov/pubmed/30082302 http://dx.doi.org/10.1098/rsta.2017.0356 |
_version_ | 1783349983406194688 |
---|---|
author | Schuemie, Martijn J. Ryan, Patrick B. Hripcsak, George Madigan, David Suchard, Marc A. |
author_facet | Schuemie, Martijn J. Ryan, Patrick B. Hripcsak, George Madigan, David Suchard, Marc A. |
author_sort | Schuemie, Martijn J. |
collection | PubMed |
description | Concerns over reproducibility in science extend to research using existing healthcare data; many observational studies investigating the same topic produce conflicting results, even when using the same data. To address this problem, we propose a paradigm shift. The current paradigm centres on generating one estimate at a time using a unique study design with unknown reliability and publishing (or not) one estimate at a time. The new paradigm advocates for high-throughput observational studies using consistent and standardized methods, allowing evaluation, calibration and unbiased dissemination to generate a more reliable and complete evidence base. We demonstrate this new paradigm by comparing all depression treatments for a set of outcomes, producing 17 718 hazard ratios, each using methodology on par with current best practice. We furthermore include control hypotheses to evaluate and calibrate our evidence generation process. Results show good transitivity and consistency between databases, and agree with four out of the five findings from clinical trials. The distribution of effect size estimates reported in the literature reveals an absence of small or null effects, with a sharp cut-off at p = 0.05. No such phenomena were observed in our results, suggesting more complete and more reliable evidence. This article is part of a discussion meeting issue ‘The growing ubiquity of algorithms in society: implications, impacts and innovations’. |
format | Online Article Text |
id | pubmed-6107542 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2018 |
publisher | The Royal Society Publishing |
record_format | MEDLINE/PubMed |
spelling | pubmed-61075422018-08-24 Improving reproducibility by using high-throughput observational studies with empirical calibration Schuemie, Martijn J. Ryan, Patrick B. Hripcsak, George Madigan, David Suchard, Marc A. Philos Trans A Math Phys Eng Sci Articles Concerns over reproducibility in science extend to research using existing healthcare data; many observational studies investigating the same topic produce conflicting results, even when using the same data. To address this problem, we propose a paradigm shift. The current paradigm centres on generating one estimate at a time using a unique study design with unknown reliability and publishing (or not) one estimate at a time. The new paradigm advocates for high-throughput observational studies using consistent and standardized methods, allowing evaluation, calibration and unbiased dissemination to generate a more reliable and complete evidence base. We demonstrate this new paradigm by comparing all depression treatments for a set of outcomes, producing 17 718 hazard ratios, each using methodology on par with current best practice. We furthermore include control hypotheses to evaluate and calibrate our evidence generation process. Results show good transitivity and consistency between databases, and agree with four out of the five findings from clinical trials. The distribution of effect size estimates reported in the literature reveals an absence of small or null effects, with a sharp cut-off at p = 0.05. No such phenomena were observed in our results, suggesting more complete and more reliable evidence. This article is part of a discussion meeting issue ‘The growing ubiquity of algorithms in society: implications, impacts and innovations’. The Royal Society Publishing 2018-09-13 2018-08-06 /pmc/articles/PMC6107542/ /pubmed/30082302 http://dx.doi.org/10.1098/rsta.2017.0356 Text en © 2018 The Authors. http://creativecommons.org/licenses/by/4.0/ Published by the Royal Society under the terms of the Creative Commons Attribution License http://creativecommons.org/licenses/by/4.0/, which permits unrestricted use, provided the original author and source are credited. |
spellingShingle | Articles Schuemie, Martijn J. Ryan, Patrick B. Hripcsak, George Madigan, David Suchard, Marc A. Improving reproducibility by using high-throughput observational studies with empirical calibration |
title | Improving reproducibility by using high-throughput observational studies with empirical calibration |
title_full | Improving reproducibility by using high-throughput observational studies with empirical calibration |
title_fullStr | Improving reproducibility by using high-throughput observational studies with empirical calibration |
title_full_unstemmed | Improving reproducibility by using high-throughput observational studies with empirical calibration |
title_short | Improving reproducibility by using high-throughput observational studies with empirical calibration |
title_sort | improving reproducibility by using high-throughput observational studies with empirical calibration |
topic | Articles |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6107542/ https://www.ncbi.nlm.nih.gov/pubmed/30082302 http://dx.doi.org/10.1098/rsta.2017.0356 |
work_keys_str_mv | AT schuemiemartijnj improvingreproducibilitybyusinghighthroughputobservationalstudieswithempiricalcalibration AT ryanpatrickb improvingreproducibilitybyusinghighthroughputobservationalstudieswithempiricalcalibration AT hripcsakgeorge improvingreproducibilitybyusinghighthroughputobservationalstudieswithempiricalcalibration AT madigandavid improvingreproducibilitybyusinghighthroughputobservationalstudieswithempiricalcalibration AT suchardmarca improvingreproducibilitybyusinghighthroughputobservationalstudieswithempiricalcalibration |