Cargando…
In Search of Ambiguity: A Three-Stage Workflow Design to Clarify Annotation Guidelines for Crowd Workers
We propose a novel three-stage FIND-RESOLVE-LABEL workflow for crowdsourced annotation to reduce ambiguity in task instructions and, thus, improve annotation quality. Stage 1 (FIND) asks the crowd to find examples whose correct label seems ambiguous given task instructions. Workers are also asked to...
Autores principales: | , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Frontiers Media S.A.
2022
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9159300/ https://www.ncbi.nlm.nih.gov/pubmed/35664506 http://dx.doi.org/10.3389/frai.2022.828187 |
_version_ | 1784719027756597248 |
---|---|
author | Pradhan, Vivek Krishna Schaekermann, Mike Lease, Matthew |
author_facet | Pradhan, Vivek Krishna Schaekermann, Mike Lease, Matthew |
author_sort | Pradhan, Vivek Krishna |
collection | PubMed |
description | We propose a novel three-stage FIND-RESOLVE-LABEL workflow for crowdsourced annotation to reduce ambiguity in task instructions and, thus, improve annotation quality. Stage 1 (FIND) asks the crowd to find examples whose correct label seems ambiguous given task instructions. Workers are also asked to provide a short tag that describes the ambiguous concept embodied by the specific instance found. We compare collaborative vs. non-collaborative designs for this stage. In Stage 2 (RESOLVE), the requester selects one or more of these ambiguous examples to label (resolving ambiguity). The new label(s) are automatically injected back into task instructions in order to improve clarity. Finally, in Stage 3 (LABEL), workers perform the actual annotation using the revised guidelines with clarifying examples. We compare three designs using these examples: examples only, tags only, or both. We report image labeling experiments over six task designs using Amazon's Mechanical Turk. Results show improved annotation accuracy and further insights regarding effective design for crowdsourced annotation tasks. |
format | Online Article Text |
id | pubmed-9159300 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2022 |
publisher | Frontiers Media S.A. |
record_format | MEDLINE/PubMed |
spelling | pubmed-91593002022-06-02 In Search of Ambiguity: A Three-Stage Workflow Design to Clarify Annotation Guidelines for Crowd Workers Pradhan, Vivek Krishna Schaekermann, Mike Lease, Matthew Front Artif Intell Artificial Intelligence We propose a novel three-stage FIND-RESOLVE-LABEL workflow for crowdsourced annotation to reduce ambiguity in task instructions and, thus, improve annotation quality. Stage 1 (FIND) asks the crowd to find examples whose correct label seems ambiguous given task instructions. Workers are also asked to provide a short tag that describes the ambiguous concept embodied by the specific instance found. We compare collaborative vs. non-collaborative designs for this stage. In Stage 2 (RESOLVE), the requester selects one or more of these ambiguous examples to label (resolving ambiguity). The new label(s) are automatically injected back into task instructions in order to improve clarity. Finally, in Stage 3 (LABEL), workers perform the actual annotation using the revised guidelines with clarifying examples. We compare three designs using these examples: examples only, tags only, or both. We report image labeling experiments over six task designs using Amazon's Mechanical Turk. Results show improved annotation accuracy and further insights regarding effective design for crowdsourced annotation tasks. Frontiers Media S.A. 2022-05-18 /pmc/articles/PMC9159300/ /pubmed/35664506 http://dx.doi.org/10.3389/frai.2022.828187 Text en Copyright © 2022 Pradhan, Schaekermann and Lease. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms. |
spellingShingle | Artificial Intelligence Pradhan, Vivek Krishna Schaekermann, Mike Lease, Matthew In Search of Ambiguity: A Three-Stage Workflow Design to Clarify Annotation Guidelines for Crowd Workers |
title | In Search of Ambiguity: A Three-Stage Workflow Design to Clarify Annotation Guidelines for Crowd Workers |
title_full | In Search of Ambiguity: A Three-Stage Workflow Design to Clarify Annotation Guidelines for Crowd Workers |
title_fullStr | In Search of Ambiguity: A Three-Stage Workflow Design to Clarify Annotation Guidelines for Crowd Workers |
title_full_unstemmed | In Search of Ambiguity: A Three-Stage Workflow Design to Clarify Annotation Guidelines for Crowd Workers |
title_short | In Search of Ambiguity: A Three-Stage Workflow Design to Clarify Annotation Guidelines for Crowd Workers |
title_sort | in search of ambiguity: a three-stage workflow design to clarify annotation guidelines for crowd workers |
topic | Artificial Intelligence |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9159300/ https://www.ncbi.nlm.nih.gov/pubmed/35664506 http://dx.doi.org/10.3389/frai.2022.828187 |
work_keys_str_mv | AT pradhanvivekkrishna insearchofambiguityathreestageworkflowdesigntoclarifyannotationguidelinesforcrowdworkers AT schaekermannmike insearchofambiguityathreestageworkflowdesigntoclarifyannotationguidelinesforcrowdworkers AT leasematthew insearchofambiguityathreestageworkflowdesigntoclarifyannotationguidelinesforcrowdworkers |