Cargando…

Introducing a Framework to Assess Newly Created Questions with Natural Language Processing

Statistical models such as those derived from Item Response Theory (IRT) enable the assessment of students on a specific subject, which can be useful for several purposes (e.g., learning path customization, drop-out prediction). However, the questions have to be assessed as well and, although it is...

Descripción completa

Detalles Bibliográficos
Autores principales: Benedetto, Luca, Cappelli, Andrea, Turrin, Roberto, Cremonesi, Paolo
Formato: Online Artículo Texto
Lenguaje:English
Publicado: 2020
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7334176/
http://dx.doi.org/10.1007/978-3-030-52237-7_4
_version_ 1783553882088013824
author Benedetto, Luca
Cappelli, Andrea
Turrin, Roberto
Cremonesi, Paolo
author_facet Benedetto, Luca
Cappelli, Andrea
Turrin, Roberto
Cremonesi, Paolo
author_sort Benedetto, Luca
collection PubMed
description Statistical models such as those derived from Item Response Theory (IRT) enable the assessment of students on a specific subject, which can be useful for several purposes (e.g., learning path customization, drop-out prediction). However, the questions have to be assessed as well and, although it is possible to estimate with IRT the characteristics of questions that have already been answered by several students, this technique cannot be used on newly generated questions. In this paper, we propose a framework to train and evaluate models for estimating the difficulty and discrimination of newly created Multiple Choice Questions by extracting meaningful features from the text of the question and of the possible choices. We implement one model using this framework and test it on a real-world dataset provided by CloudAcademy, showing that it outperforms previously proposed models, reducing by 6.7% the RMSE for difficulty estimation and by 10.8% the RMSE for discrimination estimation. We also present the results of an ablation study performed to support our features choice and to show the effects of different characteristics of the questions’ text on difficulty and discrimination.
format Online
Article
Text
id pubmed-7334176
institution National Center for Biotechnology Information
language English
publishDate 2020
record_format MEDLINE/PubMed
spelling pubmed-73341762020-07-06 Introducing a Framework to Assess Newly Created Questions with Natural Language Processing Benedetto, Luca Cappelli, Andrea Turrin, Roberto Cremonesi, Paolo Artificial Intelligence in Education Article Statistical models such as those derived from Item Response Theory (IRT) enable the assessment of students on a specific subject, which can be useful for several purposes (e.g., learning path customization, drop-out prediction). However, the questions have to be assessed as well and, although it is possible to estimate with IRT the characteristics of questions that have already been answered by several students, this technique cannot be used on newly generated questions. In this paper, we propose a framework to train and evaluate models for estimating the difficulty and discrimination of newly created Multiple Choice Questions by extracting meaningful features from the text of the question and of the possible choices. We implement one model using this framework and test it on a real-world dataset provided by CloudAcademy, showing that it outperforms previously proposed models, reducing by 6.7% the RMSE for difficulty estimation and by 10.8% the RMSE for discrimination estimation. We also present the results of an ablation study performed to support our features choice and to show the effects of different characteristics of the questions’ text on difficulty and discrimination. 2020-06-09 /pmc/articles/PMC7334176/ http://dx.doi.org/10.1007/978-3-030-52237-7_4 Text en © Springer Nature Switzerland AG 2020 This article is made available via the PMC Open Access Subset for unrestricted research re-use and secondary analysis in any form or by any means with acknowledgement of the original source. These permissions are granted for the duration of the World Health Organization (WHO) declaration of COVID-19 as a global pandemic.
spellingShingle Article
Benedetto, Luca
Cappelli, Andrea
Turrin, Roberto
Cremonesi, Paolo
Introducing a Framework to Assess Newly Created Questions with Natural Language Processing
title Introducing a Framework to Assess Newly Created Questions with Natural Language Processing
title_full Introducing a Framework to Assess Newly Created Questions with Natural Language Processing
title_fullStr Introducing a Framework to Assess Newly Created Questions with Natural Language Processing
title_full_unstemmed Introducing a Framework to Assess Newly Created Questions with Natural Language Processing
title_short Introducing a Framework to Assess Newly Created Questions with Natural Language Processing
title_sort introducing a framework to assess newly created questions with natural language processing
topic Article
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7334176/
http://dx.doi.org/10.1007/978-3-030-52237-7_4
work_keys_str_mv AT benedettoluca introducingaframeworktoassessnewlycreatedquestionswithnaturallanguageprocessing
AT cappelliandrea introducingaframeworktoassessnewlycreatedquestionswithnaturallanguageprocessing
AT turrinroberto introducingaframeworktoassessnewlycreatedquestionswithnaturallanguageprocessing
AT cremonesipaolo introducingaframeworktoassessnewlycreatedquestionswithnaturallanguageprocessing