Cargando…
Robust Neural Automated Essay Scoring Using Item Response Theory
Automated essay scoring (AES) is the task of automatically assigning scores to essays as an alternative to human grading. Conventional AES methods typically rely on manually tuned features, which are laborious to effectively develop. To obviate the need for feature engineering, many deep neural netw...
Autores principales: | , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
2020
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7334153/ http://dx.doi.org/10.1007/978-3-030-52237-7_44 |
_version_ | 1783553876133150720 |
---|---|
author | Uto, Masaki Okano, Masashi |
author_facet | Uto, Masaki Okano, Masashi |
author_sort | Uto, Masaki |
collection | PubMed |
description | Automated essay scoring (AES) is the task of automatically assigning scores to essays as an alternative to human grading. Conventional AES methods typically rely on manually tuned features, which are laborious to effectively develop. To obviate the need for feature engineering, many deep neural network (DNN)-based AES models have been proposed and have achieved state-of-the-art accuracy. DNN-AES models require training on a large dataset of graded essays. However, assigned grades in such datasets are known to be strongly biased due to effects of rater bias when grading is conducted by assigning a few raters in a rater set to each essay. Performance of DNN models rapidly drops when such biased data are used for model training. In the fields of educational and psychological measurement, item response theory (IRT) models that can estimate essay scores while considering effects of rater characteristics have recently been proposed. This study therefore proposes a new DNN-AES framework that integrates IRT models to deal with rater bias within training data. To our knowledge, this is a first attempt at addressing rating bias effects in training data, which is a crucial but overlooked problem. |
format | Online Article Text |
id | pubmed-7334153 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2020 |
record_format | MEDLINE/PubMed |
spelling | pubmed-73341532020-07-06 Robust Neural Automated Essay Scoring Using Item Response Theory Uto, Masaki Okano, Masashi Artificial Intelligence in Education Article Automated essay scoring (AES) is the task of automatically assigning scores to essays as an alternative to human grading. Conventional AES methods typically rely on manually tuned features, which are laborious to effectively develop. To obviate the need for feature engineering, many deep neural network (DNN)-based AES models have been proposed and have achieved state-of-the-art accuracy. DNN-AES models require training on a large dataset of graded essays. However, assigned grades in such datasets are known to be strongly biased due to effects of rater bias when grading is conducted by assigning a few raters in a rater set to each essay. Performance of DNN models rapidly drops when such biased data are used for model training. In the fields of educational and psychological measurement, item response theory (IRT) models that can estimate essay scores while considering effects of rater characteristics have recently been proposed. This study therefore proposes a new DNN-AES framework that integrates IRT models to deal with rater bias within training data. To our knowledge, this is a first attempt at addressing rating bias effects in training data, which is a crucial but overlooked problem. 2020-06-09 /pmc/articles/PMC7334153/ http://dx.doi.org/10.1007/978-3-030-52237-7_44 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic. |
spellingShingle | Article Uto, Masaki Okano, Masashi Robust Neural Automated Essay Scoring Using Item Response Theory |
title | Robust Neural Automated Essay Scoring Using Item Response Theory |
title_full | Robust Neural Automated Essay Scoring Using Item Response Theory |
title_fullStr | Robust Neural Automated Essay Scoring Using Item Response Theory |
title_full_unstemmed | Robust Neural Automated Essay Scoring Using Item Response Theory |
title_short | Robust Neural Automated Essay Scoring Using Item Response Theory |
title_sort | robust neural automated essay scoring using item response theory |
topic | Article |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7334153/ http://dx.doi.org/10.1007/978-3-030-52237-7_44 |
work_keys_str_mv | AT utomasaki robustneuralautomatedessayscoringusingitemresponsetheory AT okanomasashi robustneuralautomatedessayscoringusingitemresponsetheory |