Cargando…
An exploration of crowdsourcing citation screening for systematic reviews
Systematic reviews are increasingly used to inform health care decisions, but are expensive to produce. We explore the use of crowdsourcing (distributing tasks to untrained workers via the web) to reduce the cost of screening citations. We used Amazon Mechanical Turk as our platform and 4 previously...
Autores principales: | , , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
John Wiley and Sons Inc.
2017
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5589498/ https://www.ncbi.nlm.nih.gov/pubmed/28677322 http://dx.doi.org/10.1002/jrsm.1252 |
_version_ | 1783262341789384704 |
---|---|
author | Mortensen, Michael L. Adam, Gaelen P. Trikalinos, Thomas A. Kraska, Tim Wallace, Byron C. |
author_facet | Mortensen, Michael L. Adam, Gaelen P. Trikalinos, Thomas A. Kraska, Tim Wallace, Byron C. |
author_sort | Mortensen, Michael L. |
collection | PubMed |
description | Systematic reviews are increasingly used to inform health care decisions, but are expensive to produce. We explore the use of crowdsourcing (distributing tasks to untrained workers via the web) to reduce the cost of screening citations. We used Amazon Mechanical Turk as our platform and 4 previously conducted systematic reviews as examples. For each citation, workers answered 4 or 5 questions that were equivalent to the eligibility criteria. We aggregated responses from multiple workers into an overall decision to include or exclude the citation using 1 of 9 algorithms and compared the performance of these algorithms to the corresponding decisions of trained experts. The most inclusive algorithm (designating a citation as relevant if any worker did) identified 95% to 99% of the citations that were ultimately included in the reviews while excluding 68% to 82% of irrelevant citations. Other algorithms increased the fraction of irrelevant articles excluded at some cost to the inclusion of relevant studies. Crowdworkers completed screening in 4 to 17 days, costing $460 to $2220, a cost reduction of up to 88% compared to trained experts. Crowdsourcing may represent a useful approach to reducing the cost of identifying literature for systematic reviews. |
format | Online Article Text |
id | pubmed-5589498 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2017 |
publisher | John Wiley and Sons Inc. |
record_format | MEDLINE/PubMed |
spelling | pubmed-55894982017-10-02 An exploration of crowdsourcing citation screening for systematic reviews Mortensen, Michael L. Adam, Gaelen P. Trikalinos, Thomas A. Kraska, Tim Wallace, Byron C. Res Synth Methods Original Articles Systematic reviews are increasingly used to inform health care decisions, but are expensive to produce. We explore the use of crowdsourcing (distributing tasks to untrained workers via the web) to reduce the cost of screening citations. We used Amazon Mechanical Turk as our platform and 4 previously conducted systematic reviews as examples. For each citation, workers answered 4 or 5 questions that were equivalent to the eligibility criteria. We aggregated responses from multiple workers into an overall decision to include or exclude the citation using 1 of 9 algorithms and compared the performance of these algorithms to the corresponding decisions of trained experts. The most inclusive algorithm (designating a citation as relevant if any worker did) identified 95% to 99% of the citations that were ultimately included in the reviews while excluding 68% to 82% of irrelevant citations. Other algorithms increased the fraction of irrelevant articles excluded at some cost to the inclusion of relevant studies. Crowdworkers completed screening in 4 to 17 days, costing $460 to $2220, a cost reduction of up to 88% compared to trained experts. Crowdsourcing may represent a useful approach to reducing the cost of identifying literature for systematic reviews. John Wiley and Sons Inc. 2017-07-04 2017-09 /pmc/articles/PMC5589498/ /pubmed/28677322 http://dx.doi.org/10.1002/jrsm.1252 Text en © 2017 The Authors. Research Synthesis Methods Published by John Wiley & Sons Ltd This is an open access article under the terms of the Creative Commons Attribution (http://creativecommons.org/licenses/by/4.0/) License, which permits use, distribution and reproduction in any medium, provided the original work is properly cited. |
spellingShingle | Original Articles Mortensen, Michael L. Adam, Gaelen P. Trikalinos, Thomas A. Kraska, Tim Wallace, Byron C. An exploration of crowdsourcing citation screening for systematic reviews |
title | An exploration of crowdsourcing citation screening for systematic reviews |
title_full | An exploration of crowdsourcing citation screening for systematic reviews |
title_fullStr | An exploration of crowdsourcing citation screening for systematic reviews |
title_full_unstemmed | An exploration of crowdsourcing citation screening for systematic reviews |
title_short | An exploration of crowdsourcing citation screening for systematic reviews |
title_sort | exploration of crowdsourcing citation screening for systematic reviews |
topic | Original Articles |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5589498/ https://www.ncbi.nlm.nih.gov/pubmed/28677322 http://dx.doi.org/10.1002/jrsm.1252 |
work_keys_str_mv | AT mortensenmichaell anexplorationofcrowdsourcingcitationscreeningforsystematicreviews AT adamgaelenp anexplorationofcrowdsourcingcitationscreeningforsystematicreviews AT trikalinosthomasa anexplorationofcrowdsourcingcitationscreeningforsystematicreviews AT kraskatim anexplorationofcrowdsourcingcitationscreeningforsystematicreviews AT wallacebyronc anexplorationofcrowdsourcingcitationscreeningforsystematicreviews AT mortensenmichaell explorationofcrowdsourcingcitationscreeningforsystematicreviews AT adamgaelenp explorationofcrowdsourcingcitationscreeningforsystematicreviews AT trikalinosthomasa explorationofcrowdsourcingcitationscreeningforsystematicreviews AT kraskatim explorationofcrowdsourcingcitationscreeningforsystematicreviews AT wallacebyronc explorationofcrowdsourcingcitationscreeningforsystematicreviews |