Evaluating the consistency of automated CEFR analyzers: a study of English language text classification
| dc.contributor.author | Siripol P. | |
| dc.contributor.author | Rhee S. | |
| dc.contributor.author | Thirakunkovit S. | |
| dc.contributor.author | Liang-Itsara A. | |
| dc.contributor.correspondence | Siripol P. | |
| dc.contributor.other | Mahidol University | |
| dc.date.accessioned | 2025-08-04T18:17:18Z | |
| dc.date.available | 2025-08-04T18:17:18Z | |
| dc.date.issued | 2025-08-01 | |
| dc.description.abstract | With the increasing use of web-based tools for text analysis, there is a growing reliance on automated systems to assess text difficulty and classify texts to the Common European Framework of Reference for Languages (CEFR). However, inconsistencies in these tools’ outputs could undermine their effectiveness for language learners and researchers. This study investigates the consistency of five widely used automated CEFR analyzer tools, including ChatGPT, by analyzing 20 English descriptive texts at CEFR levels B1 and B2. A quantitative approach was employed to compare the CEFR classifications generated by these tools. The results reveal significant inconsistencies across the tools, raising concerns about the reliability of automated CEFR alignment. Additionally, the content and genre of texts appeared to influence the CEFR classification, suggesting that certain factors beyond the tools’ algorithms may affect their accuracy. These findings have important implications for language educators, curriculum designers, and researchers who rely on automated CEFR tools for text selection, grading, and analysis. The study highlights the limitations of automated CEFR classification systems and calls for a more qualitative approach to text difficulty alignment analysis. Future research recommendation is discussed and call for more focus on refining these tools and exploring additional factors that may impact their effectiveness in text difficulty measurement and CEFR alignment. | |
| dc.identifier.citation | International Journal of Evaluation and Research in Education Vol.14 No.4 (2025) , 3283-3294 | |
| dc.identifier.doi | 10.11591/ijere.v14i4.33528 | |
| dc.identifier.eissn | 26205440 | |
| dc.identifier.issn | 22528822 | |
| dc.identifier.scopus | 2-s2.0-105011988557 | |
| dc.identifier.uri | https://repository.li.mahidol.ac.th/handle/123456789/111533 | |
| dc.rights.holder | SCOPUS | |
| dc.subject | Social Sciences | |
| dc.title | Evaluating the consistency of automated CEFR analyzers: a study of English language text classification | |
| dc.type | Article | |
| mu.datasource.scopus | https://www.scopus.com/inward/record.uri?partnerID=HzOxMe3b&scp=105011988557&origin=inward | |
| oaire.citation.endPage | 3294 | |
| oaire.citation.issue | 4 | |
| oaire.citation.startPage | 3283 | |
| oaire.citation.title | International Journal of Evaluation and Research in Education | |
| oaire.citation.volume | 14 | |
| oairecerif.author.affiliation | Mahidol University |
