The analysis of marking reliability through the approach of gauge repeatability and reproducibility (GR&R) study: a case of English-speaking test

dc.contributor.authorSureeyatanapas P.
dc.contributor.authorSureeyatanapas P.
dc.contributor.authorPanitanarak U.
dc.contributor.authorKraisriwattana J.
dc.contributor.authorSarootyanapat P.
dc.contributor.authorO’Connell D.
dc.contributor.correspondenceSureeyatanapas P.
dc.contributor.otherMahidol University
dc.date.accessioned2024-02-08T18:14:15Z
dc.date.available2024-02-08T18:14:15Z
dc.date.issued2024-12-01
dc.description.abstractEnsuring consistent and reliable scoring is paramount in education, especially in performance-based assessments. This study delves into the critical issue of marking consistency, focusing on speaking proficiency tests in English language learning, which often face greater reliability challenges. While existing literature has explored various methods for assessing marking reliability, this study is the first of its kind to introduce an alternative statistical tool, namely the gauge repeatability and reproducibility (GR&R) approach, to the educational context. The study encompasses both intra- and inter-rater reliabilities, with additional validation using the intraclass correlation coefficient (ICC). Using a case study approach involving three examiners evaluating 30 recordings of a speaking proficiency test, the GR&R method demonstrates its effectiveness in detecting reliability issues over the ICC approach. Furthermore, this research identifies key factors influencing scoring inconsistencies, including group performance estimation, work presentation order, rubric complexity and clarity, the student’s chosen topic, accent familiarity, and recording quality. Importantly, it not only pinpoints these root causes but also suggests practical solutions, thereby enhancing the precision of the measurement system. The GR&R method can offer significant contributions to stakeholders in language proficiency assessment, including educational institutions, test developers and policymakers. It is also applicable to other cases of performance-based assessments. By addressing reliability issues, this study provides insights to enhance the fairness and accuracy of subjective judgements, ultimately benefiting overall performance comparisons and decision making.
dc.identifier.citationLanguage Testing in Asia Vol.14 No.1 (2024)
dc.identifier.doi10.1186/s40468-023-00271-z
dc.identifier.eissn22290443
dc.identifier.scopus2-s2.0-85182491063
dc.identifier.urihttps://repository.li.mahidol.ac.th/handle/123456789/95798
dc.rights.holderSCOPUS
dc.subjectSocial Sciences
dc.subjectArts and Humanities
dc.titleThe analysis of marking reliability through the approach of gauge repeatability and reproducibility (GR&R) study: a case of English-speaking test
dc.typeArticle
mu.datasource.scopushttps://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=85182491063&origin=inward
oaire.citation.issue1
oaire.citation.titleLanguage Testing in Asia
oaire.citation.volume14
oairecerif.author.affiliationKhon Kaen University International College
oairecerif.author.affiliationKhon Kaen University
oairecerif.author.affiliationMahidol University

Files

Collections