Cargando…

Peer Review Evaluation Process of Marie Curie Actions under EU’s Seventh Framework Programme for Research

We analysed the peer review of grant proposals under Marie Curie Actions, a major EU research funding instrument, which involves two steps: an independent assessment (Individual Evaluation Report, IER) performed remotely by 3 raters, and a consensus opinion reached during a meeting by the same rater...

Descripción completa

Detalles Bibliográficos
Autores principales: Pina, David G., Hren, Darko, Marušić, Ana
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Public Library of Science 2015
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4488366/
https://www.ncbi.nlm.nih.gov/pubmed/26126111
http://dx.doi.org/10.1371/journal.pone.0130753
_version_ 1782379143024869376
author Pina, David G.
Hren, Darko
Marušić, Ana
author_facet Pina, David G.
Hren, Darko
Marušić, Ana
author_sort Pina, David G.
collection PubMed
description We analysed the peer review of grant proposals under Marie Curie Actions, a major EU research funding instrument, which involves two steps: an independent assessment (Individual Evaluation Report, IER) performed remotely by 3 raters, and a consensus opinion reached during a meeting by the same raters (Consensus Report, CR). For 24,897 proposals evaluated from 2007 to 2013, the association between average IER and CR scores was very high across different panels, grant calls and years. Median average deviation (AD) index, used as a measure of inter-rater agreement, was 5.4 points on a 0-100 scale (interquartile range 3.4-8.3), overall, demonstrating a good general agreement among raters. For proposals where one rater disagreed with the other two raters (n=1424; 5.7%), or where all 3 raters disagreed (n=2075; 8.3%), the average IER and CR scores were still highly associated. Disagreement was more frequent for proposals from Economics/Social Sciences and Humanities panels. Greater disagreement was observed for proposals with lower average IER scores. CR scores for proposals with initial disagreement were also significantly lower. Proposals with a large absolute difference between the average IER and CR scores (≥10 points; n=368, 1.5%) generally had lower CR scores. An inter-correlation matrix of individual raters' scores of evaluation criteria of proposals indicated that these scores were, in general, a reflection of raters’ overall scores. Our analysis demonstrated a good internal consistency and general high agreement among raters. Consensus meetings appear to be relevant for particular panels and subsets of proposals with large differences among raters’ scores.
format Online
Article
Text
id pubmed-4488366
institution National Center for Biotechnology Information
language English
publishDate 2015
publisher Public Library of Science
record_format MEDLINE/PubMed
spelling pubmed-44883662015-07-02 Peer Review Evaluation Process of Marie Curie Actions under EU’s Seventh Framework Programme for Research Pina, David G. Hren, Darko Marušić, Ana PLoS One Research Article We analysed the peer review of grant proposals under Marie Curie Actions, a major EU research funding instrument, which involves two steps: an independent assessment (Individual Evaluation Report, IER) performed remotely by 3 raters, and a consensus opinion reached during a meeting by the same raters (Consensus Report, CR). For 24,897 proposals evaluated from 2007 to 2013, the association between average IER and CR scores was very high across different panels, grant calls and years. Median average deviation (AD) index, used as a measure of inter-rater agreement, was 5.4 points on a 0-100 scale (interquartile range 3.4-8.3), overall, demonstrating a good general agreement among raters. For proposals where one rater disagreed with the other two raters (n=1424; 5.7%), or where all 3 raters disagreed (n=2075; 8.3%), the average IER and CR scores were still highly associated. Disagreement was more frequent for proposals from Economics/Social Sciences and Humanities panels. Greater disagreement was observed for proposals with lower average IER scores. CR scores for proposals with initial disagreement were also significantly lower. Proposals with a large absolute difference between the average IER and CR scores (≥10 points; n=368, 1.5%) generally had lower CR scores. An inter-correlation matrix of individual raters' scores of evaluation criteria of proposals indicated that these scores were, in general, a reflection of raters’ overall scores. Our analysis demonstrated a good internal consistency and general high agreement among raters. Consensus meetings appear to be relevant for particular panels and subsets of proposals with large differences among raters’ scores. Public Library of Science 2015-06-30 /pmc/articles/PMC4488366/ /pubmed/26126111 http://dx.doi.org/10.1371/journal.pone.0130753 Text en © 2015 Pina et al http://creativecommons.org/licenses/by/4.0/ This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are properly credited.
spellingShingle Research Article
Pina, David G.
Hren, Darko
Marušić, Ana
Peer Review Evaluation Process of Marie Curie Actions under EU’s Seventh Framework Programme for Research
title Peer Review Evaluation Process of Marie Curie Actions under EU’s Seventh Framework Programme for Research
title_full Peer Review Evaluation Process of Marie Curie Actions under EU’s Seventh Framework Programme for Research
title_fullStr Peer Review Evaluation Process of Marie Curie Actions under EU’s Seventh Framework Programme for Research
title_full_unstemmed Peer Review Evaluation Process of Marie Curie Actions under EU’s Seventh Framework Programme for Research
title_short Peer Review Evaluation Process of Marie Curie Actions under EU’s Seventh Framework Programme for Research
title_sort peer review evaluation process of marie curie actions under eu’s seventh framework programme for research
topic Research Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC4488366/
https://www.ncbi.nlm.nih.gov/pubmed/26126111
http://dx.doi.org/10.1371/journal.pone.0130753
work_keys_str_mv AT pinadavidg peerreviewevaluationprocessofmariecurieactionsundereusseventhframeworkprogrammeforresearch
AT hrendarko peerreviewevaluationprocessofmariecurieactionsundereusseventhframeworkprogrammeforresearch
AT marusicana peerreviewevaluationprocessofmariecurieactionsundereusseventhframeworkprogrammeforresearch