Cargando…
GradeAid: a framework for automatic short answers grading in educational contexts—design, implementation and evaluation
Automatic short answer grading (ASAG), a hot field of natural language understanding, is a research area within learning analytics. ASAG solutions are conceived to offload teachers and instructors, especially those in higher education, where classes with hundreds of students are the norm and the tas...
Autores principales: | , , , |
---|---|
Formato: | Online Artículo Texto |
Lenguaje: | English |
Publicado: |
Springer London
2023
|
Materias: | |
Acceso en línea: | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10197042/ https://www.ncbi.nlm.nih.gov/pubmed/37361374 http://dx.doi.org/10.1007/s10115-023-01892-9 |
_version_ | 1785044468837122048 |
---|---|
author | del Gobbo, Emiliano Guarino, Alfonso Cafarelli, Barbara Grilli, Luca |
author_facet | del Gobbo, Emiliano Guarino, Alfonso Cafarelli, Barbara Grilli, Luca |
author_sort | del Gobbo, Emiliano |
collection | PubMed |
description | Automatic short answer grading (ASAG), a hot field of natural language understanding, is a research area within learning analytics. ASAG solutions are conceived to offload teachers and instructors, especially those in higher education, where classes with hundreds of students are the norm and the task of grading (short)answers to open-ended questionnaires becomes tougher. Their outcomes are precious both for the very grading and for providing students with “ad hoc” feedback. ASAG proposals have also enabled different intelligent tutoring systems. Over the years, a variety of ASAG solutions have been proposed, still there are a series of gaps in the literature that we fill in this paper. The present work proposes GradeAid, a framework for ASAG. It is based on the joint analysis of lexical and semantic features of the students’ answers through state-of-the-art regressors; differently from any other previous work, (i) it copes with non-English datasets, (ii) it has undergone a robust validation and benchmarking phase, and (iii) it has been tested on every dataset publicly available and on a new dataset (now available for researchers). GradeAid obtains performance comparable to the systems presented in the literature (root-mean-squared errors down to 0.25 based on the specific tuple [Formula: see text] dataset-question[Formula: see text] ). We argue it represents a strong baseline for further developments in the field. |
format | Online Article Text |
id | pubmed-10197042 |
institution | National Center for Biotechnology Information |
language | English |
publishDate | 2023 |
publisher | Springer London |
record_format | MEDLINE/PubMed |
spelling | pubmed-101970422023-05-23 GradeAid: a framework for automatic short answers grading in educational contexts—design, implementation and evaluation del Gobbo, Emiliano Guarino, Alfonso Cafarelli, Barbara Grilli, Luca Knowl Inf Syst Regular Paper Automatic short answer grading (ASAG), a hot field of natural language understanding, is a research area within learning analytics. ASAG solutions are conceived to offload teachers and instructors, especially those in higher education, where classes with hundreds of students are the norm and the task of grading (short)answers to open-ended questionnaires becomes tougher. Their outcomes are precious both for the very grading and for providing students with “ad hoc” feedback. ASAG proposals have also enabled different intelligent tutoring systems. Over the years, a variety of ASAG solutions have been proposed, still there are a series of gaps in the literature that we fill in this paper. The present work proposes GradeAid, a framework for ASAG. It is based on the joint analysis of lexical and semantic features of the students’ answers through state-of-the-art regressors; differently from any other previous work, (i) it copes with non-English datasets, (ii) it has undergone a robust validation and benchmarking phase, and (iii) it has been tested on every dataset publicly available and on a new dataset (now available for researchers). GradeAid obtains performance comparable to the systems presented in the literature (root-mean-squared errors down to 0.25 based on the specific tuple [Formula: see text] dataset-question[Formula: see text] ). We argue it represents a strong baseline for further developments in the field. Springer London 2023-05-19 /pmc/articles/PMC10197042/ /pubmed/37361374 http://dx.doi.org/10.1007/s10115-023-01892-9 Text en © The Author(s) 2023 https://creativecommons.org/licenses/by/4.0/Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/ (https://creativecommons.org/licenses/by/4.0/) . |
spellingShingle | Regular Paper del Gobbo, Emiliano Guarino, Alfonso Cafarelli, Barbara Grilli, Luca GradeAid: a framework for automatic short answers grading in educational contexts—design, implementation and evaluation |
title | GradeAid: a framework for automatic short answers grading in educational contexts—design, implementation and evaluation |
title_full | GradeAid: a framework for automatic short answers grading in educational contexts—design, implementation and evaluation |
title_fullStr | GradeAid: a framework for automatic short answers grading in educational contexts—design, implementation and evaluation |
title_full_unstemmed | GradeAid: a framework for automatic short answers grading in educational contexts—design, implementation and evaluation |
title_short | GradeAid: a framework for automatic short answers grading in educational contexts—design, implementation and evaluation |
title_sort | gradeaid: a framework for automatic short answers grading in educational contexts—design, implementation and evaluation |
topic | Regular Paper |
url | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10197042/ https://www.ncbi.nlm.nih.gov/pubmed/37361374 http://dx.doi.org/10.1007/s10115-023-01892-9 |
work_keys_str_mv | AT delgobboemiliano gradeaidaframeworkforautomaticshortanswersgradingineducationalcontextsdesignimplementationandevaluation AT guarinoalfonso gradeaidaframeworkforautomaticshortanswersgradingineducationalcontextsdesignimplementationandevaluation AT cafarellibarbara gradeaidaframeworkforautomaticshortanswersgradingineducationalcontextsdesignimplementationandevaluation AT grilliluca gradeaidaframeworkforautomaticshortanswersgradingineducationalcontextsdesignimplementationandevaluation |