Cargando…
Metrics and methods in the evaluation of prestige bias in peer review: A case study in computer systems conferences
The integrity of peer review is essential for modern science. Numerous studies have therefore focused on identifying, quantifying, and mitigating biases in peer review. One of these better-known biases is prestige bias, where the recognition of a famous author or affiliation leads reviewers to subco...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8880855/ https://www.ncbi.nlm.nih.gov/pubmed/35213600 http://dx.doi.org/10.1371/journal.pone.0264131 |
_version_ | 1784659324931407872 |
---|---|
author | Frachtenberg, Eitan McConville, Kelly S. |
author_facet | Frachtenberg, Eitan McConville, Kelly S. |
author_sort | Frachtenberg, Eitan |
collection | PubMed |
description | The integrity of peer review is essential for modern science. Numerous studies have therefore focused on identifying, quantifying, and mitigating biases in peer review. One of these better-known biases is prestige bias, where the recognition of a famous author or affiliation leads reviewers to subconsciously treat their submissions preferentially. A common mitigation approach for prestige bias is double-blind reviewing, where the identify of authors is hidden from reviewers. However, studies on the effectivness of this mitigation are mixed and are rarely directly comparable to each other, leading to difficulty in generalization of their results. In this paper, we explore the design space for such studies in an attempt to reach common ground. Using an observational approach with a large dataset of peer-reviewed papers in computer systems, we systematically evaluate the effects of different prestige metrics, aggregation methods, control variables, and outlier treatments. We show that depending on these choices, the data can lead to contradictory conclusions with high statistical significance. For example, authors with higher h-index often preferred to publish in competitive conferences which are also typically double-blind, whereas authors with higher paper counts often preferred the single-blind conferences. The main practical implication of our analyses is that a narrow evaluation may lead to unreliable results. A thorough evaluation of prestige bias requires a careful inventory of assumptions, metrics, and methodology, often requiring a more detailed sensitivity analysis than is normally undertaken. Importantly, two of the most commonly used metrics for prestige evaluation, past publication count and h-index, are not independent from the choice of publishing venue, which must be accounted for when comparing authors prestige across conferences. |
format | Online Article Text |
id | pubmed-8880855 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-88808552022-02-26 Metrics and methods in the evaluation of prestige bias in peer review: A case study in computer systems conferences Frachtenberg, Eitan McConville, Kelly S. PLoS One Research Article The integrity of peer review is essential for modern science. Numerous studies have therefore focused on identifying, quantifying, and mitigating biases in peer review. One of these better-known biases is prestige bias, where the recognition of a famous author or affiliation leads reviewers to subconsciously treat their submissions preferentially. A common mitigation approach for prestige bias is double-blind reviewing, where the identify of authors is hidden from reviewers. However, studies on the effectivness of this mitigation are mixed and are rarely directly comparable to each other, leading to difficulty in generalization of their results. In this paper, we explore the design space for such studies in an attempt to reach common ground. Using an observational approach with a large dataset of peer-reviewed papers in computer systems, we systematically evaluate the effects of different prestige metrics, aggregation methods, control variables, and outlier treatments. We show that depending on these choices, the data can lead to contradictory conclusions with high statistical significance. For example, authors with higher h-index often preferred to publish in competitive conferences which are also typically double-blind, whereas authors with higher paper counts often preferred the single-blind conferences. The main practical implication of our analyses is that a narrow evaluation may lead to unreliable results. A thorough evaluation of prestige bias requires a careful inventory of assumptions, metrics, and methodology, often requiring a more detailed sensitivity analysis than is normally undertaken. Importantly, two of the most commonly used metrics for prestige evaluation, past publication count and h-index, are not independent from the choice of publishing venue, which must be accounted for when comparing authors prestige across conferences. Public Library of Science 2022-02-25 /pmc/articles/PMC8880855/ /pubmed/35213600 http://dx.doi.org/10.1371/journal.pone.0264131 Text en © 2022 Frachtenberg, McConville https://creativecommons.org/licenses/by/4.0/This is an open access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/) , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited. |
spellingShingle | Research Article Frachtenberg, Eitan McConville, Kelly S. Metrics and methods in the evaluation of prestige bias in peer review: A case study in computer systems conferences |
title | Metrics and methods in the evaluation of prestige bias in peer review: A case study in computer systems conferences |
title_full | Metrics and methods in the evaluation of prestige bias in peer review: A case study in computer systems conferences |
title_fullStr | Metrics and methods in the evaluation of prestige bias in peer review: A case study in computer systems conferences |
title_full_unstemmed | Metrics and methods in the evaluation of prestige bias in peer review: A case study in computer systems conferences |
title_short | Metrics and methods in the evaluation of prestige bias in peer review: A case study in computer systems conferences |
title_sort | metrics and methods in the evaluation of prestige bias in peer review: a case study in computer systems conferences |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8880855/ https://www.ncbi.nlm.nih.gov/pubmed/35213600 http://dx.doi.org/10.1371/journal.pone.0264131 |
work_keys_str_mv | AT frachtenbergeitan metricsandmethodsintheevaluationofprestigebiasinpeerreviewacasestudyincomputersystemsconferences AT mcconvillekellys metricsandmethodsintheevaluationofprestigebiasinpeerreviewacasestudyincomputersystemsconferences |