Evaluating the consistency of automated CEFR analyzers: a study of English language text classification

dc.contributor.authorSiripol P.
dc.contributor.authorRhee S.
dc.contributor.authorThirakunkovit S.
dc.contributor.authorLiang-Itsara A.
dc.contributor.correspondenceSiripol P.
dc.contributor.otherMahidol University
dc.date.accessioned2025-08-04T18:17:18Z
dc.date.available2025-08-04T18:17:18Z
dc.date.issued2025-08-01
dc.description.abstractWith the increasing use of web-based tools for text analysis, there is a growing reliance on automated systems to assess text difficulty and classify texts to the Common European Framework of Reference for Languages (CEFR). However, inconsistencies in these tools’ outputs could undermine their effectiveness for language learners and researchers. This study investigates the consistency of five widely used automated CEFR analyzer tools, including ChatGPT, by analyzing 20 English descriptive texts at CEFR levels B1 and B2. A quantitative approach was employed to compare the CEFR classifications generated by these tools. The results reveal significant inconsistencies across the tools, raising concerns about the reliability of automated CEFR alignment. Additionally, the content and genre of texts appeared to influence the CEFR classification, suggesting that certain factors beyond the tools’ algorithms may affect their accuracy. These findings have important implications for language educators, curriculum designers, and researchers who rely on automated CEFR tools for text selection, grading, and analysis. The study highlights the limitations of automated CEFR classification systems and calls for a more qualitative approach to text difficulty alignment analysis. Future research recommendation is discussed and call for more focus on refining these tools and exploring additional factors that may impact their effectiveness in text difficulty measurement and CEFR alignment.
dc.identifier.citationInternational Journal of Evaluation and Research in Education Vol.14 No.4 (2025) , 3283-3294
dc.identifier.doi10.11591/ijere.v14i4.33528
dc.identifier.eissn26205440
dc.identifier.issn22528822
dc.identifier.scopus2-s2.0-105011988557
dc.identifier.urihttps://repository.li.mahidol.ac.th/handle/123456789/111533
dc.rights.holderSCOPUS
dc.subjectSocial Sciences
dc.titleEvaluating the consistency of automated CEFR analyzers: a study of English language text classification
dc.typeArticle
mu.datasource.scopushttps://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=105011988557&origin=inward
oaire.citation.endPage3294
oaire.citation.issue4
oaire.citation.startPage3283
oaire.citation.titleInternational Journal of Evaluation and Research in Education
oaire.citation.volume14
oairecerif.author.affiliationMahidol University

Files

Collections