Cargando…

An exploration of crowdsourcing citation screening for systematic reviews

Systematic reviews are increasingly used to inform health care decisions, but are expensive to produce. We explore the use of crowdsourcing (distributing tasks to untrained workers via the web) to reduce the cost of screening citations. We used Amazon Mechanical Turk as our platform and 4 previously...

Descripción completa

Detalles Bibliográficos
Autores principales: Mortensen, Michael L., Adam, Gaelen P., Trikalinos, Thomas A., Kraska, Tim, Wallace, Byron C.
Formato: Online Artículo Texto
Lenguaje:English
Publicado: John Wiley and Sons Inc. 2017
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5589498/
https://www.ncbi.nlm.nih.gov/pubmed/28677322
http://dx.doi.org/10.1002/jrsm.1252
_version_ 1783262341789384704
author Mortensen, Michael L.
Adam, Gaelen P.
Trikalinos, Thomas A.
Kraska, Tim
Wallace, Byron C.
author_facet Mortensen, Michael L.
Adam, Gaelen P.
Trikalinos, Thomas A.
Kraska, Tim
Wallace, Byron C.
author_sort Mortensen, Michael L.
collection PubMed
description Systematic reviews are increasingly used to inform health care decisions, but are expensive to produce. We explore the use of crowdsourcing (distributing tasks to untrained workers via the web) to reduce the cost of screening citations. We used Amazon Mechanical Turk as our platform and 4 previously conducted systematic reviews as examples. For each citation, workers answered 4 or 5 questions that were equivalent to the eligibility criteria. We aggregated responses from multiple workers into an overall decision to include or exclude the citation using 1 of 9 algorithms and compared the performance of these algorithms to the corresponding decisions of trained experts. The most inclusive algorithm (designating a citation as relevant if any worker did) identified 95% to 99% of the citations that were ultimately included in the reviews while excluding 68% to 82% of irrelevant citations. Other algorithms increased the fraction of irrelevant articles excluded at some cost to the inclusion of relevant studies. Crowdworkers completed screening in 4 to 17 days, costing $460 to $2220, a cost reduction of up to 88% compared to trained experts. Crowdsourcing may represent a useful approach to reducing the cost of identifying literature for systematic reviews.
format Online
Article
Text
id pubmed-5589498
institution National Center for Biotechnology Information
language English
publishDate 2017
publisher John Wiley and Sons Inc.
record_format MEDLINE/PubMed
spelling pubmed-55894982017-10-02 An exploration of crowdsourcing citation screening for systematic reviews Mortensen, Michael L. Adam, Gaelen P. Trikalinos, Thomas A. Kraska, Tim Wallace, Byron C. Res Synth Methods Original Articles Systematic reviews are increasingly used to inform health care decisions, but are expensive to produce. We explore the use of crowdsourcing (distributing tasks to untrained workers via the web) to reduce the cost of screening citations. We used Amazon Mechanical Turk as our platform and 4 previously conducted systematic reviews as examples. For each citation, workers answered 4 or 5 questions that were equivalent to the eligibility criteria. We aggregated responses from multiple workers into an overall decision to include or exclude the citation using 1 of 9 algorithms and compared the performance of these algorithms to the corresponding decisions of trained experts. The most inclusive algorithm (designating a citation as relevant if any worker did) identified 95% to 99% of the citations that were ultimately included in the reviews while excluding 68% to 82% of irrelevant citations. Other algorithms increased the fraction of irrelevant articles excluded at some cost to the inclusion of relevant studies. Crowdworkers completed screening in 4 to 17 days, costing $460 to $2220, a cost reduction of up to 88% compared to trained experts. Crowdsourcing may represent a useful approach to reducing the cost of identifying literature for systematic reviews. John Wiley and Sons Inc. 2017-07-04 2017-09 /pmc/articles/PMC5589498/ /pubmed/28677322 http://dx.doi.org/10.1002/jrsm.1252 Text en © 2017 The Authors. Research Synthesis Methods Published by John Wiley & Sons Ltd This is an open access article under the terms of the Creative Commons Attribution (http://creativecommons.org/licenses/by/4.0/) License, which permits use, distribution and reproduction in any medium, provided the original work is properly cited.
spellingShingle Original Articles
Mortensen, Michael L.
Adam, Gaelen P.
Trikalinos, Thomas A.
Kraska, Tim
Wallace, Byron C.
An exploration of crowdsourcing citation screening for systematic reviews
title An exploration of crowdsourcing citation screening for systematic reviews
title_full An exploration of crowdsourcing citation screening for systematic reviews
title_fullStr An exploration of crowdsourcing citation screening for systematic reviews
title_full_unstemmed An exploration of crowdsourcing citation screening for systematic reviews
title_short An exploration of crowdsourcing citation screening for systematic reviews
title_sort exploration of crowdsourcing citation screening for systematic reviews
topic Original Articles
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5589498/
https://www.ncbi.nlm.nih.gov/pubmed/28677322
http://dx.doi.org/10.1002/jrsm.1252
work_keys_str_mv AT mortensenmichaell anexplorationofcrowdsourcingcitationscreeningforsystematicreviews
AT adamgaelenp anexplorationofcrowdsourcingcitationscreeningforsystematicreviews
AT trikalinosthomasa anexplorationofcrowdsourcingcitationscreeningforsystematicreviews
AT kraskatim anexplorationofcrowdsourcingcitationscreeningforsystematicreviews
AT wallacebyronc anexplorationofcrowdsourcingcitationscreeningforsystematicreviews
AT mortensenmichaell explorationofcrowdsourcingcitationscreeningforsystematicreviews
AT adamgaelenp explorationofcrowdsourcingcitationscreeningforsystematicreviews
AT trikalinosthomasa explorationofcrowdsourcingcitationscreeningforsystematicreviews
AT kraskatim explorationofcrowdsourcingcitationscreeningforsystematicreviews
AT wallacebyronc explorationofcrowdsourcingcitationscreeningforsystematicreviews