Cargando…
How to Get the Most out of Your Curation Effort
Large-scale annotation efforts typically involve several experts who may disagree with each other. We propose an approach for modeling disagreements among experts that allows providing each annotation with a confidence value (i.e., the posterior probability that it is correct). Our approach allows c...
Autores principales: | , , |
---|---|
Formato: | Texto |
Lenguaje: | English |
Publicado: |
Public Library of Science
2009
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2678295/ https://www.ncbi.nlm.nih.gov/pubmed/19461884 http://dx.doi.org/10.1371/journal.pcbi.1000391 |
_version_ | 1782166842556547072 |
---|---|
author | Rzhetsky, Andrey Shatkay, Hagit Wilbur, W. John |
author_facet | Rzhetsky, Andrey Shatkay, Hagit Wilbur, W. John |
author_sort | Rzhetsky, Andrey |
collection | PubMed |
description | Large-scale annotation efforts typically involve several experts who may disagree with each other. We propose an approach for modeling disagreements among experts that allows providing each annotation with a confidence value (i.e., the posterior probability that it is correct). Our approach allows computing certainty-level for individual annotations, given annotator-specific parameters estimated from data. We developed two probabilistic models for performing this analysis, compared these models using computer simulation, and tested each model's actual performance, based on a large data set generated by human annotators specifically for this study. We show that even in the worst-case scenario, when all annotators disagree, our approach allows us to significantly increase the probability of choosing the correct annotation. Along with this publication we make publicly available a corpus of 10,000 sentences annotated according to several cardinal dimensions that we have introduced in earlier work. The 10,000 sentences were all 3-fold annotated by a group of eight experts, while a 1,000-sentence subset was further 5-fold annotated by five new experts. While the presented data represent a specialized curation task, our modeling approach is general; most data annotation studies could benefit from our methodology. |
format | Text |
id | pubmed-2678295 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2009 |
publisher | Public Library of Science |
record_format | MEDLINE/PubMed |
spelling | pubmed-26782952009-05-22 How to Get the Most out of Your Curation Effort Rzhetsky, Andrey Shatkay, Hagit Wilbur, W. John PLoS Comput Biol Research Article Large-scale annotation efforts typically involve several experts who may disagree with each other. We propose an approach for modeling disagreements among experts that allows providing each annotation with a confidence value (i.e., the posterior probability that it is correct). Our approach allows computing certainty-level for individual annotations, given annotator-specific parameters estimated from data. We developed two probabilistic models for performing this analysis, compared these models using computer simulation, and tested each model's actual performance, based on a large data set generated by human annotators specifically for this study. We show that even in the worst-case scenario, when all annotators disagree, our approach allows us to significantly increase the probability of choosing the correct annotation. Along with this publication we make publicly available a corpus of 10,000 sentences annotated according to several cardinal dimensions that we have introduced in earlier work. The 10,000 sentences were all 3-fold annotated by a group of eight experts, while a 1,000-sentence subset was further 5-fold annotated by five new experts. While the presented data represent a specialized curation task, our modeling approach is general; most data annotation studies could benefit from our methodology. Public Library of Science 2009-05-22 /pmc/articles/PMC2678295/ /pubmed/19461884 http://dx.doi.org/10.1371/journal.pcbi.1000391 Text en This is an open-access article distributed under the terms of the Creative Commons Public Domain declaration which stipulates that, once placed in the public domain, this work may be freely reproduced, distributed, transmitted, modified, built upon, or otherwise used by anyone for any lawful purpose. https://creativecommons.org/publicdomain/zero/1.0/ This is an open-access article distributed under the terms of the Creative Commons Public Domain declaration, which stipulates that, once placed in the public domain, this work may be freely reproduced, distributed, transmitted, modified, built upon, or otherwise used by anyone for any lawful purpose. |
spellingShingle | Research Article Rzhetsky, Andrey Shatkay, Hagit Wilbur, W. John How to Get the Most out of Your Curation Effort |
title | How to Get the Most out of Your Curation Effort |
title_full | How to Get the Most out of Your Curation Effort |
title_fullStr | How to Get the Most out of Your Curation Effort |
title_full_unstemmed | How to Get the Most out of Your Curation Effort |
title_short | How to Get the Most out of Your Curation Effort |
title_sort | how to get the most out of your curation effort |
topic | Research Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC2678295/ https://www.ncbi.nlm.nih.gov/pubmed/19461884 http://dx.doi.org/10.1371/journal.pcbi.1000391 |
work_keys_str_mv | AT rzhetskyandrey howtogetthemostoutofyourcurationeffort AT shatkayhagit howtogetthemostoutofyourcurationeffort AT wilburwjohn howtogetthemostoutofyourcurationeffort |