Cargando…

Robust Neural Automated Essay Scoring Using Item Response Theory

Automated essay scoring (AES) is the task of automatically assigning scores to essays as an alternative to human grading. Conventional AES methods typically rely on manually tuned features, which are laborious to effectively develop. To obviate the need for feature engineering, many deep neural netw...

Descripción completa

Detalles Bibliográficos
Autores principales: Uto, Masaki, Okano, Masashi
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7334153/
http://dx.doi.org/10.1007/978-3-030-52237-7_44
_version_ 1783553876133150720
author Uto, Masaki
Okano, Masashi
author_facet Uto, Masaki
Okano, Masashi
author_sort Uto, Masaki
collection PubMed
description Automated essay scoring (AES) is the task of automatically assigning scores to essays as an alternative to human grading. Conventional AES methods typically rely on manually tuned features, which are laborious to effectively develop. To obviate the need for feature engineering, many deep neural network (DNN)-based AES models have been proposed and have achieved state-of-the-art accuracy. DNN-AES models require training on a large dataset of graded essays. However, assigned grades in such datasets are known to be strongly biased due to effects of rater bias when grading is conducted by assigning a few raters in a rater set to each essay. Performance of DNN models rapidly drops when such biased data are used for model training. In the fields of educational and psychological measurement, item response theory (IRT) models that can estimate essay scores while considering effects of rater characteristics have recently been proposed. This study therefore proposes a new DNN-AES framework that integrates IRT models to deal with rater bias within training data. To our knowledge, this is a first attempt at addressing rating bias effects in training data, which is a crucial but overlooked problem.
format Online
Article
Text
id pubmed-7334153
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-73341532020-07-06 Robust Neural Automated Essay Scoring Using Item Response Theory Uto, Masaki Okano, Masashi Artificial Intelligence in Education Article Automated essay scoring (AES) is the task of automatically assigning scores to essays as an alternative to human grading. Conventional AES methods typically rely on manually tuned features, which are laborious to effectively develop. To obviate the need for feature engineering, many deep neural network (DNN)-based AES models have been proposed and have achieved state-of-the-art accuracy. DNN-AES models require training on a large dataset of graded essays. However, assigned grades in such datasets are known to be strongly biased due to effects of rater bias when grading is conducted by assigning a few raters in a rater set to each essay. Performance of DNN models rapidly drops when such biased data are used for model training. In the fields of educational and psychological measurement, item response theory (IRT) models that can estimate essay scores while considering effects of rater characteristics have recently been proposed. This study therefore proposes a new DNN-AES framework that integrates IRT models to deal with rater bias within training data. To our knowledge, this is a first attempt at addressing rating bias effects in training data, which is a crucial but overlooked problem. 2020-06-09 /pmc/articles/PMC7334153/ http://dx.doi.org/10.1007/978-3-030-52237-7_44 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Uto, Masaki
Okano, Masashi
Robust Neural Automated Essay Scoring Using Item Response Theory
title Robust Neural Automated Essay Scoring Using Item Response Theory
title_full Robust Neural Automated Essay Scoring Using Item Response Theory
title_fullStr Robust Neural Automated Essay Scoring Using Item Response Theory
title_full_unstemmed Robust Neural Automated Essay Scoring Using Item Response Theory
title_short Robust Neural Automated Essay Scoring Using Item Response Theory
title_sort robust neural automated essay scoring using item response theory
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7334153/
http://dx.doi.org/10.1007/978-3-030-52237-7_44
work_keys_str_mv AT utomasaki robustneuralautomatedessayscoringusingitemresponsetheory
AT okanomasashi robustneuralautomatedessayscoringusingitemresponsetheory