Cargando…

The raters’ differences in Arabic writing rubrics through the Many-Facet Rasch measurement model

Writing assessment relies closely on scoring the excellence of a subject’s thoughts. This creates a faceted measurement structure regarding rubrics, tasks, and raters. Nevertheless, most studies did not consider the differences among raters systematically. This study examines the raters’ differences...

Descripción completa

Detalles Bibliográficos
Autores principales: Baharudin, Harun, Maskor, Zunita Mohamad, Matore, Mohd Effendi Ewan Mohd
Formato: Online Artículo Texto
Lenguaje:English
Publicado: Frontiers Media S.A. 2022
Materias:
Acceso en línea:https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9801977/
https://www.ncbi.nlm.nih.gov/pubmed/36591072
http://dx.doi.org/10.3389/fpsyg.2022.988272
_version_ 1784861598135877632
author Baharudin, Harun
Maskor, Zunita Mohamad
Matore, Mohd Effendi Ewan Mohd
author_facet Baharudin, Harun
Maskor, Zunita Mohamad
Matore, Mohd Effendi Ewan Mohd
author_sort Baharudin, Harun
collection PubMed
description Writing assessment relies closely on scoring the excellence of a subject’s thoughts. This creates a faceted measurement structure regarding rubrics, tasks, and raters. Nevertheless, most studies did not consider the differences among raters systematically. This study examines the raters’ differences in association with the reliability and validity of writing rubrics using the Many-Facet Rasch measurement model (MFRM) to model these differences. A set of standards for evaluating the quality of rating based on writing assessment was examined. Rating quality was tested within four writing domains from an analytic rubric using a scale of one to three. The writing domains explored were vocabulary, grammar, language, use, and organization; whereas the data were obtained from 15 Arabic essays gathered from religious secondary school students under the supervision of the Malaysia Ministry of Education. Five raters in the field of practice were selected to evaluate all the essays. As a result, (a) raters range considerably on the lenient-severity dimension, so rater variations ought to be modeled; (b) the combination of findings between raters avoids the doubt of scores, thereby reducing the measurement error which could lower the criterion validity with the external variable; and (c) MFRM adjustments effectively increased the correlations of the scores obtained from partial and full data. Predominant findings revealed that rating quality varies across analytic rubric domains. This also depicts that MFRM is an effective way to model rater differences and evaluate the validity and reliability of writing rubrics.
format Online
Article
Text
id pubmed-9801977
institution National Center for Biotechnology Information
language English
publishDate 2022
publisher Frontiers Media S.A.
record_format MEDLINE/PubMed
spelling pubmed-98019772022-12-31 The raters’ differences in Arabic writing rubrics through the Many-Facet Rasch measurement model Baharudin, Harun Maskor, Zunita Mohamad Matore, Mohd Effendi Ewan Mohd Front Psychol Psychology Writing assessment relies closely on scoring the excellence of a subject’s thoughts. This creates a faceted measurement structure regarding rubrics, tasks, and raters. Nevertheless, most studies did not consider the differences among raters systematically. This study examines the raters’ differences in association with the reliability and validity of writing rubrics using the Many-Facet Rasch measurement model (MFRM) to model these differences. A set of standards for evaluating the quality of rating based on writing assessment was examined. Rating quality was tested within four writing domains from an analytic rubric using a scale of one to three. The writing domains explored were vocabulary, grammar, language, use, and organization; whereas the data were obtained from 15 Arabic essays gathered from religious secondary school students under the supervision of the Malaysia Ministry of Education. Five raters in the field of practice were selected to evaluate all the essays. As a result, (a) raters range considerably on the lenient-severity dimension, so rater variations ought to be modeled; (b) the combination of findings between raters avoids the doubt of scores, thereby reducing the measurement error which could lower the criterion validity with the external variable; and (c) MFRM adjustments effectively increased the correlations of the scores obtained from partial and full data. Predominant findings revealed that rating quality varies across analytic rubric domains. This also depicts that MFRM is an effective way to model rater differences and evaluate the validity and reliability of writing rubrics. Frontiers Media S.A. 2022-12-16 /pmc/articles/PMC9801977/ /pubmed/36591072 http://dx.doi.org/10.3389/fpsyg.2022.988272 Text en Copyright © 2022 Baharudin, Maskor and Matore. https://creativecommons.org/licenses/by/4.0/This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
spellingShingle Psychology
Baharudin, Harun
Maskor, Zunita Mohamad
Matore, Mohd Effendi Ewan Mohd
The raters’ differences in Arabic writing rubrics through the Many-Facet Rasch measurement model
title The raters’ differences in Arabic writing rubrics through the Many-Facet Rasch measurement model
title_full The raters’ differences in Arabic writing rubrics through the Many-Facet Rasch measurement model
title_fullStr The raters’ differences in Arabic writing rubrics through the Many-Facet Rasch measurement model
title_full_unstemmed The raters’ differences in Arabic writing rubrics through the Many-Facet Rasch measurement model
title_short The raters’ differences in Arabic writing rubrics through the Many-Facet Rasch measurement model
title_sort raters’ differences in arabic writing rubrics through the many-facet rasch measurement model
topic Psychology
url https://www.ncbi.nlm.nih.gov/pmc/articles/PMC9801977/
https://www.ncbi.nlm.nih.gov/pubmed/36591072
http://dx.doi.org/10.3389/fpsyg.2022.988272
work_keys_str_mv AT baharudinharun theratersdifferencesinarabicwritingrubricsthroughthemanyfacetraschmeasurementmodel
AT maskorzunitamohamad theratersdifferencesinarabicwritingrubricsthroughthemanyfacetraschmeasurementmodel
AT matoremohdeffendiewanmohd theratersdifferencesinarabicwritingrubricsthroughthemanyfacetraschmeasurementmodel
AT baharudinharun ratersdifferencesinarabicwritingrubricsthroughthemanyfacetraschmeasurementmodel
AT maskorzunitamohamad ratersdifferencesinarabicwritingrubricsthroughthemanyfacetraschmeasurementmodel
AT matoremohdeffendiewanmohd ratersdifferencesinarabicwritingrubricsthroughthemanyfacetraschmeasurementmodel