Cargando…

Error rates of human reviewers during abstract screening in systematic reviews

BACKGROUND: Automated approaches to improve the efficiency of systematic reviews are greatly needed. When testing any of these approaches, the criterion standard of comparison (gold standard) is usually human reviewers. Yet, human reviewers make errors in inclusion and exclusion of references. OBJEC...

Descripción completa

Detalles Bibliográficos
Autores principales: Wang, Zhen, Nayfeh, Tarek, Tetzlaff, Jennifer, O’Blenis, Peter, Murad, Mohammad Hassan
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6959565/
https://www.ncbi.nlm.nih.gov/pubmed/31935267
http://dx.doi.org/10.1371/journal.pone.0227742
_version_ 1783487613080961024
author Wang, Zhen
Nayfeh, Tarek
Tetzlaff, Jennifer
O’Blenis, Peter
Murad, Mohammad Hassan
author_facet Wang, Zhen
Nayfeh, Tarek
Tetzlaff, Jennifer
O’Blenis, Peter
Murad, Mohammad Hassan
author_sort Wang, Zhen
collection PubMed
description BACKGROUND: Automated approaches to improve the efficiency of systematic reviews are greatly needed. When testing any of these approaches, the criterion standard of comparison (gold standard) is usually human reviewers. Yet, human reviewers make errors in inclusion and exclusion of references. OBJECTIVES: To determine citation false inclusion and false exclusion rates during abstract screening by pairs of independent reviewers. These rates can help in designing, testing and implementing automated approaches. METHODS: We identified all systematic reviews conducted between 2010 and 2017 by an evidence-based practice center in the United States. Eligible reviews had to follow standard systematic review procedures with dual independent screening of abstracts and full texts, in which citation inclusion by one reviewer prompted automatic inclusion through the next level of screening. Disagreements between reviewers during full text screening were reconciled via consensus or arbitration by a third reviewer. A false inclusion or exclusion was defined as a decision made by a single reviewer that was inconsistent with the final included list of studies. RESULTS: We analyzed a total of 139,467 citations that underwent 329,332 inclusion and exclusion decisions from 86 unique reviewers. The final systematic reviews included 5.48% of the potential references identified through bibliographic database search (95% confidence interval (CI): 2.38% to 8.58%). After abstract screening, the total error rate (false inclusion and false exclusion) was 10.76% (95% CI: 7.43% to 14.09%). CONCLUSIONS: This study suggests important false inclusion and exclusion rates by human reviewers. When deciding the validity of a future automated study selection algorithm, it is important to keep in mind that the gold standard is not perfect and that achieving error rates similar to humans may be adequate and can save resources and time.
format Online
Article
Text
id pubmed-6959565
institution National Center for Biotechnology Information
language English
publishDate 2020
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-69595652020-01-26 Error rates of human reviewers during abstract screening in systematic reviews Wang, Zhen Nayfeh, Tarek Tetzlaff, Jennifer O’Blenis, Peter Murad, Mohammad Hassan PLoS One Research Article BACKGROUND: Automated approaches to improve the efficiency of systematic reviews are greatly needed. When testing any of these approaches, the criterion standard of comparison (gold standard) is usually human reviewers. Yet, human reviewers make errors in inclusion and exclusion of references. OBJECTIVES: To determine citation false inclusion and false exclusion rates during abstract screening by pairs of independent reviewers. These rates can help in designing, testing and implementing automated approaches. METHODS: We identified all systematic reviews conducted between 2010 and 2017 by an evidence-based practice center in the United States. Eligible reviews had to follow standard systematic review procedures with dual independent screening of abstracts and full texts, in which citation inclusion by one reviewer prompted automatic inclusion through the next level of screening. Disagreements between reviewers during full text screening were reconciled via consensus or arbitration by a third reviewer. A false inclusion or exclusion was defined as a decision made by a single reviewer that was inconsistent with the final included list of studies. RESULTS: We analyzed a total of 139,467 citations that underwent 329,332 inclusion and exclusion decisions from 86 unique reviewers. The final systematic reviews included 5.48% of the potential references identified through bibliographic database search (95% confidence interval (CI): 2.38% to 8.58%). After abstract screening, the total error rate (false inclusion and false exclusion) was 10.76% (95% CI: 7.43% to 14.09%). CONCLUSIONS: This study suggests important false inclusion and exclusion rates by human reviewers. When deciding the validity of a future automated study selection algorithm, it is important to keep in mind that the gold standard is not perfect and that achieving error rates similar to humans may be adequate and can save resources and time. Public Library of Science 2020-01-14 /pmc/articles/PMC6959565/ /pubmed/31935267 http://dx.doi.org/10.1371/journal.pone.0227742 Text en © 2020 Wang et al http://creativecommons.org/licenses/by/4.0/ This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
spellingShingle Research Article
Wang, Zhen
Nayfeh, Tarek
Tetzlaff, Jennifer
O’Blenis, Peter
Murad, Mohammad Hassan
Error rates of human reviewers during abstract screening in systematic reviews
title Error rates of human reviewers during abstract screening in systematic reviews
title_full Error rates of human reviewers during abstract screening in systematic reviews
title_fullStr Error rates of human reviewers during abstract screening in systematic reviews
title_full_unstemmed Error rates of human reviewers during abstract screening in systematic reviews
title_short Error rates of human reviewers during abstract screening in systematic reviews
title_sort error rates of human reviewers during abstract screening in systematic reviews
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6959565/
https://www.ncbi.nlm.nih.gov/pubmed/31935267
http://dx.doi.org/10.1371/journal.pone.0227742
work_keys_str_mv AT wangzhen errorratesofhumanreviewersduringabstractscreeninginsystematicreviews
AT nayfehtarek errorratesofhumanreviewersduringabstractscreeninginsystematicreviews
AT tetzlaffjennifer errorratesofhumanreviewersduringabstractscreeninginsystematicreviews
AT oblenispeter errorratesofhumanreviewersduringabstractscreeninginsystematicreviews
AT muradmohammadhassan errorratesofhumanreviewersduringabstractscreeninginsystematicreviews