Cargando…

Utility of Crowdsourced User Experiments for Measuring the Central Tendency of User Performance: A Case of Error-Rate Model Evaluation in a Pointing Task

The usage of crowdsourcing to recruit numerous participants has been recognized as beneficial in the human-computer interaction (HCI) field, such as for designing user interfaces and validating user performance models. In this work, we investigate its effectiveness for evaluating an error-rate predi...

Descripción completa

Detalles Bibliográficos
Autor principal: Yamanaka, Shota
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8993509/
https://www.ncbi.nlm.nih.gov/pubmed/35402900
http://dx.doi.org/10.3389/frai.2022.798892
_version_ 1784683915443699712
author Yamanaka, Shota
author_facet Yamanaka, Shota
author_sort Yamanaka, Shota
collection PubMed
description The usage of crowdsourcing to recruit numerous participants has been recognized as beneficial in the human-computer interaction (HCI) field, such as for designing user interfaces and validating user performance models. In this work, we investigate its effectiveness for evaluating an error-rate prediction model in target pointing tasks. In contrast to models for operational times, a clicking error (i.e., missing a target) occurs by chance at a certain probability, e.g., 5%. Therefore, in traditional laboratory-based experiments, a lot of repetitions are needed to measure the central tendency of error rates. We hypothesize that recruiting many workers would enable us to keep the number of repetitions per worker much smaller. We collected data from 384 workers and found that existing models on operational time and error rate showed good fits (both R(2) > 0.95). A simulation where we changed the number of participants N(P) and the number of repetitions N(repeat) showed that the time prediction model was robust against small N(P) and N(repeat), although the error-rate model fitness was considerably degraded. These findings empirically demonstrate a new utility of crowdsourced user experiments for collecting numerous participants, which should be of great use to HCI researchers for their evaluation studies.
format Online
Article
Text
id pubmed-8993509
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-89935092022-04-09 Utility of Crowdsourced User Experiments for Measuring the Central Tendency of User Performance: A Case of Error-Rate Model Evaluation in a Pointing Task Yamanaka, Shota Front Artif Intell Artificial Intelligence The usage of crowdsourcing to recruit numerous participants has been recognized as beneficial in the human-computer interaction (HCI) field, such as for designing user interfaces and validating user performance models. In this work, we investigate its effectiveness for evaluating an error-rate prediction model in target pointing tasks. In contrast to models for operational times, a clicking error (i.e., missing a target) occurs by chance at a certain probability, e.g., 5%. Therefore, in traditional laboratory-based experiments, a lot of repetitions are needed to measure the central tendency of error rates. We hypothesize that recruiting many workers would enable us to keep the number of repetitions per worker much smaller. We collected data from 384 workers and found that existing models on operational time and error rate showed good fits (both R(2) > 0.95). A simulation where we changed the number of participants N(P) and the number of repetitions N(repeat) showed that the time prediction model was robust against small N(P) and N(repeat), although the error-rate model fitness was considerably degraded. These findings empirically demonstrate a new utility of crowdsourced user experiments for collecting numerous participants, which should be of great use to HCI researchers for their evaluation studies. Frontiers Media S.A. 2022-03-17 /pmc/articles/PMC8993509/ /pubmed/35402900 http://dx.doi.org/10.3389/frai.2022.798892 Text en Copyright © 2022 Yamanaka. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Artificial Intelligence
Yamanaka, Shota
Utility of Crowdsourced User Experiments for Measuring the Central Tendency of User Performance: A Case of Error-Rate Model Evaluation in a Pointing Task
title Utility of Crowdsourced User Experiments for Measuring the Central Tendency of User Performance: A Case of Error-Rate Model Evaluation in a Pointing Task
title_full Utility of Crowdsourced User Experiments for Measuring the Central Tendency of User Performance: A Case of Error-Rate Model Evaluation in a Pointing Task
title_fullStr Utility of Crowdsourced User Experiments for Measuring the Central Tendency of User Performance: A Case of Error-Rate Model Evaluation in a Pointing Task
title_full_unstemmed Utility of Crowdsourced User Experiments for Measuring the Central Tendency of User Performance: A Case of Error-Rate Model Evaluation in a Pointing Task
title_short Utility of Crowdsourced User Experiments for Measuring the Central Tendency of User Performance: A Case of Error-Rate Model Evaluation in a Pointing Task
title_sort utility of crowdsourced user experiments for measuring the central tendency of user performance: a case of error-rate model evaluation in a pointing task
topic Artificial Intelligence
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC8993509/
https://www.ncbi.nlm.nih.gov/pubmed/35402900
http://dx.doi.org/10.3389/frai.2022.798892
work_keys_str_mv AT yamanakashota utilityofcrowdsourceduserexperimentsformeasuringthecentraltendencyofuserperformanceacaseoferrorratemodelevaluationinapointingtask