Cargando…

Automating risk of bias assessment in systematic reviews: a real-time mixed methods comparison of human researchers to a machine learning system

BACKGROUND: Machine learning and automation are increasingly used to make the evidence synthesis process faster and more responsive to policymakers’ needs. In systematic reviews of randomized controlled trials (RCTs), risk of bias assessment is a resource-intensive task that typically requires two t...

Descripción completa

Detalles Bibliográficos
Autores principales: Jardim, Patricia Sofia Jacobsen, Rose, Christopher James, Ames, Heather Melanie, Echavez, Jose Francisco Meneses, Van de Velde, Stijn, Muller, Ashley Elizabeth
Formato: Online Artículo Texto
Lenguaje:English
Publicado: BioMed Central 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9174024/
https://www.ncbi.nlm.nih.gov/pubmed/35676632
http://dx.doi.org/10.1186/s12874-022-01649-y
_version_ 1784722149141905408
author Jardim, Patricia Sofia Jacobsen
Rose, Christopher James
Ames, Heather Melanie
Echavez, Jose Francisco Meneses
Van de Velde, Stijn
Muller, Ashley Elizabeth
author_facet Jardim, Patricia Sofia Jacobsen
Rose, Christopher James
Ames, Heather Melanie
Echavez, Jose Francisco Meneses
Van de Velde, Stijn
Muller, Ashley Elizabeth
author_sort Jardim, Patricia Sofia Jacobsen
collection PubMed
description BACKGROUND: Machine learning and automation are increasingly used to make the evidence synthesis process faster and more responsive to policymakers’ needs. In systematic reviews of randomized controlled trials (RCTs), risk of bias assessment is a resource-intensive task that typically requires two trained reviewers. One function of RobotReviewer, an off-the-shelf machine learning system, is an automated risk of bias assessment. METHODS: We assessed the feasibility of adopting RobotReviewer within a national public health institute using a randomized, real-time, user-centered study. The study included 26 RCTs and six reviewers from two projects examining health and social interventions. We randomized these studies to one of two RobotReviewer platforms. We operationalized feasibility as accuracy, time use, and reviewer acceptability. We measured accuracy by the number of corrections made by human reviewers (either to automated assessments or another human reviewer’s assessments). We explored acceptability through group discussions and individual email responses after presenting the quantitative results. RESULTS: Reviewers were equally likely to accept judgment by RobotReviewer as each other’s judgement during the consensus process when measured dichotomously; risk ratio 1.02 (95% CI 0.92 to 1.13; p = 0.33). We were not able to compare time use. The acceptability of the program by researchers was mixed. Less experienced reviewers were generally more positive, and they saw more benefits and were able to use the tool more flexibly. Reviewers positioned human input and human-to-human interaction as superior to even a semi-automation of this process. CONCLUSION: Despite being presented with evidence of RobotReviewer’s equal performance to humans, participating reviewers were not interested in modifying standard procedures to include automation. If further studies confirm equal accuracy and reduced time compared to manual practices, we suggest that the benefits of RobotReviewer may support its future implementation as one of two assessors, despite reviewer ambivalence. Future research should study barriers to adopting automated tools and how highly educated and experienced researchers can adapt to a job market that is increasingly challenged by new technologies. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1186/s12874-022-01649-y.
format Online
Article
Text
id pubmed-9174024
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher BioMed Central
record_format MEDLINE/PubMed
spelling pubmed-91740242022-06-08 Automating risk of bias assessment in systematic reviews: a real-time mixed methods comparison of human researchers to a machine learning system Jardim, Patricia Sofia Jacobsen Rose, Christopher James Ames, Heather Melanie Echavez, Jose Francisco Meneses Van de Velde, Stijn Muller, Ashley Elizabeth BMC Med Res Methodol Research BACKGROUND: Machine learning and automation are increasingly used to make the evidence synthesis process faster and more responsive to policymakers’ needs. In systematic reviews of randomized controlled trials (RCTs), risk of bias assessment is a resource-intensive task that typically requires two trained reviewers. One function of RobotReviewer, an off-the-shelf machine learning system, is an automated risk of bias assessment. METHODS: We assessed the feasibility of adopting RobotReviewer within a national public health institute using a randomized, real-time, user-centered study. The study included 26 RCTs and six reviewers from two projects examining health and social interventions. We randomized these studies to one of two RobotReviewer platforms. We operationalized feasibility as accuracy, time use, and reviewer acceptability. We measured accuracy by the number of corrections made by human reviewers (either to automated assessments or another human reviewer’s assessments). We explored acceptability through group discussions and individual email responses after presenting the quantitative results. RESULTS: Reviewers were equally likely to accept judgment by RobotReviewer as each other’s judgement during the consensus process when measured dichotomously; risk ratio 1.02 (95% CI 0.92 to 1.13; p = 0.33). We were not able to compare time use. The acceptability of the program by researchers was mixed. Less experienced reviewers were generally more positive, and they saw more benefits and were able to use the tool more flexibly. Reviewers positioned human input and human-to-human interaction as superior to even a semi-automation of this process. CONCLUSION: Despite being presented with evidence of RobotReviewer’s equal performance to humans, participating reviewers were not interested in modifying standard procedures to include automation. If further studies confirm equal accuracy and reduced time compared to manual practices, we suggest that the benefits of RobotReviewer may support its future implementation as one of two assessors, despite reviewer ambivalence. Future research should study barriers to adopting automated tools and how highly educated and experienced researchers can adapt to a job market that is increasingly challenged by new technologies. SUPPLEMENTARY INFORMATION: The online version contains supplementary material available at 10.1186/s12874-022-01649-y. BioMed Central 2022-06-08 /pmc/articles/PMC9174024/ /pubmed/35676632 http://dx.doi.org/10.1186/s12874-022-01649-y Text en © The Author(s) 2022 https://creativecommons.org/licenses/by/4.0/Open AccessThis article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/ (https://creativecommons.org/publicdomain/zero/1.0/) ) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
spellingShingle Research
Jardim, Patricia Sofia Jacobsen
Rose, Christopher James
Ames, Heather Melanie
Echavez, Jose Francisco Meneses
Van de Velde, Stijn
Muller, Ashley Elizabeth
Automating risk of bias assessment in systematic reviews: a real-time mixed methods comparison of human researchers to a machine learning system
title Automating risk of bias assessment in systematic reviews: a real-time mixed methods comparison of human researchers to a machine learning system
title_full Automating risk of bias assessment in systematic reviews: a real-time mixed methods comparison of human researchers to a machine learning system
title_fullStr Automating risk of bias assessment in systematic reviews: a real-time mixed methods comparison of human researchers to a machine learning system
title_full_unstemmed Automating risk of bias assessment in systematic reviews: a real-time mixed methods comparison of human researchers to a machine learning system
title_short Automating risk of bias assessment in systematic reviews: a real-time mixed methods comparison of human researchers to a machine learning system
title_sort automating risk of bias assessment in systematic reviews: a real-time mixed methods comparison of human researchers to a machine learning system
topic Research
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9174024/
https://www.ncbi.nlm.nih.gov/pubmed/35676632
http://dx.doi.org/10.1186/s12874-022-01649-y
work_keys_str_mv AT jardimpatriciasofiajacobsen automatingriskofbiasassessmentinsystematicreviewsarealtimemixedmethodscomparisonofhumanresearcherstoamachinelearningsystem
AT rosechristopherjames automatingriskofbiasassessmentinsystematicreviewsarealtimemixedmethodscomparisonofhumanresearcherstoamachinelearningsystem
AT amesheathermelanie automatingriskofbiasassessmentinsystematicreviewsarealtimemixedmethodscomparisonofhumanresearcherstoamachinelearningsystem
AT echavezjosefranciscomeneses automatingriskofbiasassessmentinsystematicreviewsarealtimemixedmethodscomparisonofhumanresearcherstoamachinelearningsystem
AT vandeveldestijn automatingriskofbiasassessmentinsystematicreviewsarealtimemixedmethodscomparisonofhumanresearcherstoamachinelearningsystem
AT mullerashleyelizabeth automatingriskofbiasassessmentinsystematicreviewsarealtimemixedmethodscomparisonofhumanresearcherstoamachinelearningsystem