Cargando…

A Comparison of Data Quality Assessment Checks in Six Data Sharing Networks

OBJECTIVE: To compare rule-based data quality (DQ) assessment approaches across multiple national clinical data sharing organizations. METHODS: Six organizations with established data quality assessment (DQA) programs provided documentation or source code describing current DQ checks. DQ checks were...

Descripción completa

Detalles Bibliográficos
Autores principales: Callahan, Tiffany J., Bauck, Alan E., Bertoch, David, Brown, Jeff, Khare, Ritu, Ryan, Patrick B., Staab, Jenny, Zozus, Meredith N., Kahn, Michael G.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Ubiquity Press 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5982846/
https://www.ncbi.nlm.nih.gov/pubmed/29881733
http://dx.doi.org/10.5334/egems.223
_version_ 1783328322427551744
author Callahan, Tiffany J.
Bauck, Alan E.
Bertoch, David
Brown, Jeff
Khare, Ritu
Ryan, Patrick B.
Staab, Jenny
Zozus, Meredith N.
Kahn, Michael G.
author_facet Callahan, Tiffany J.
Bauck, Alan E.
Bertoch, David
Brown, Jeff
Khare, Ritu
Ryan, Patrick B.
Staab, Jenny
Zozus, Meredith N.
Kahn, Michael G.
author_sort Callahan, Tiffany J.
collection PubMed
description OBJECTIVE: To compare rule-based data quality (DQ) assessment approaches across multiple national clinical data sharing organizations. METHODS: Six organizations with established data quality assessment (DQA) programs provided documentation or source code describing current DQ checks. DQ checks were mapped to the categories within the data verification context of the harmonized DQA terminology. To ensure all DQ checks were consistently mapped, conventions were developed and four iterations of mapping performed. Difficult-to-map DQ checks were discussed with research team members until consensus was achieved. RESULTS: Participating organizations provided 11,026 DQ checks, of which 99.97 percent were successfully mapped to a DQA category. Of the mapped DQ checks (N=11,023), 214 (1.94 percent) mapped to multiple DQA categories. The majority of DQ checks mapped to Atemporal Plausibility (49.60 percent), Value Conformance (17.84 percent), and Atemporal Completeness (12.98 percent) categories. DISCUSSION: Using the common DQA terminology, near-complete (99.97 percent) coverage across a wide range of DQA programs and specifications was reached. Comparing the distributions of mapped DQ checks revealed important differences between participating organizations. This variation may be related to the organization’s stakeholder requirements, primary analytical focus, or maturity of their DQA program. Not within scope, mapping checks within the data validation context of the terminology may provide additional insights into DQA practice differences. CONCLUSION: A common DQA terminology provides a means to help organizations and researchers understand the coverage of their current DQA efforts as well as highlight potential areas for additional DQA development. Sharing DQ checks between organizations could help expand the scope of DQA across clinical data networks.
format Online
Article
Text
id pubmed-5982846
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher Ubiquity Press
record_format MEDLINE/PubMed
spelling pubmed-59828462018-06-07 A Comparison of Data Quality Assessment Checks in Six Data Sharing Networks Callahan, Tiffany J. Bauck, Alan E. Bertoch, David Brown, Jeff Khare, Ritu Ryan, Patrick B. Staab, Jenny Zozus, Meredith N. Kahn, Michael G. EGEMS (Wash DC) Research OBJECTIVE: To compare rule-based data quality (DQ) assessment approaches across multiple national clinical data sharing organizations. METHODS: Six organizations with established data quality assessment (DQA) programs provided documentation or source code describing current DQ checks. DQ checks were mapped to the categories within the data verification context of the harmonized DQA terminology. To ensure all DQ checks were consistently mapped, conventions were developed and four iterations of mapping performed. Difficult-to-map DQ checks were discussed with research team members until consensus was achieved. RESULTS: Participating organizations provided 11,026 DQ checks, of which 99.97 percent were successfully mapped to a DQA category. Of the mapped DQ checks (N=11,023), 214 (1.94 percent) mapped to multiple DQA categories. The majority of DQ checks mapped to Atemporal Plausibility (49.60 percent), Value Conformance (17.84 percent), and Atemporal Completeness (12.98 percent) categories. DISCUSSION: Using the common DQA terminology, near-complete (99.97 percent) coverage across a wide range of DQA programs and specifications was reached. Comparing the distributions of mapped DQ checks revealed important differences between participating organizations. This variation may be related to the organization’s stakeholder requirements, primary analytical focus, or maturity of their DQA program. Not within scope, mapping checks within the data validation context of the terminology may provide additional insights into DQA practice differences. CONCLUSION: A common DQA terminology provides a means to help organizations and researchers understand the coverage of their current DQA efforts as well as highlight potential areas for additional DQA development. Sharing DQ checks between organizations could help expand the scope of DQA across clinical data networks. Ubiquity Press 2017-06-12 /pmc/articles/PMC5982846/ /pubmed/29881733 http://dx.doi.org/10.5334/egems.223 Text en Copyright: © 2018 The Author(s) https://creativecommons.org/licenses/by-nc-nd/3.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution-NonCommercial-NoDerivs 3.0 Unported (CC BY-NC-ND 3.0), which permits unrestricted use and distribution, for non-commercial purposes, as long as the original material has not been modified, and provided the original author and source are credited. See https://creativecommons.org/licenses/by-nc-nd/3.0/.
spellingShingle Research
Callahan, Tiffany J.
Bauck, Alan E.
Bertoch, David
Brown, Jeff
Khare, Ritu
Ryan, Patrick B.
Staab, Jenny
Zozus, Meredith N.
Kahn, Michael G.
A Comparison of Data Quality Assessment Checks in Six Data Sharing Networks
title A Comparison of Data Quality Assessment Checks in Six Data Sharing Networks
title_full A Comparison of Data Quality Assessment Checks in Six Data Sharing Networks
title_fullStr A Comparison of Data Quality Assessment Checks in Six Data Sharing Networks
title_full_unstemmed A Comparison of Data Quality Assessment Checks in Six Data Sharing Networks
title_short A Comparison of Data Quality Assessment Checks in Six Data Sharing Networks
title_sort comparison of data quality assessment checks in six data sharing networks
topic Research
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5982846/
https://www.ncbi.nlm.nih.gov/pubmed/29881733
http://dx.doi.org/10.5334/egems.223
work_keys_str_mv AT callahantiffanyj acomparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT bauckalane acomparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT bertochdavid acomparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT brownjeff acomparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT khareritu acomparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT ryanpatrickb acomparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT staabjenny acomparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT zozusmeredithn acomparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT kahnmichaelg acomparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT callahantiffanyj comparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT bauckalane comparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT bertochdavid comparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT brownjeff comparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT khareritu comparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT ryanpatrickb comparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT staabjenny comparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT zozusmeredithn comparisonofdataqualityassessmentchecksinsixdatasharingnetworks
AT kahnmichaelg comparisonofdataqualityassessmentchecksinsixdatasharingnetworks