Publication in conference proceedings
UniversalCEFR: Enabling Open Multilingual Research on Language Proficiency Assessment
Joseph Marvin Imperial (Imperial, J.M.); Abdullah Barayan (Barayan, A.); Regina Stodden (Stodden, R.); Rodrigo Wilkens (Wilkens, R.); Ricardo Muñoz Sánchez (Muñoz Sánchez, R.); Lingyun Gao (Gao, L.); Melissa Torgbi (Torgbi, M.); Dawn Knight (Knight, D.); Gail Forey (Forey, G.); Reka R. Jablonkai (Jablonkai, R.R.); Ekaterina Kochmar (Kochmar, E.); Robert Reynolds (Reynolds, R.); Eugénio Ribeiro (Ribeiro, E.); Horacio Saggion (Saggion, H.); Elena Volodina (Volodina, E.); Sowmya Vajjala (Vajjala, S.); Thomas François (François, T.); Fernando Alva-Manchego (Alva-Manchego, F.); Harish Tayyar Madabushi (Tayyar Madabushi, H.); et al.
Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing
Year (definitive publication)
2025
Language
English
Country
China
More Information
Web of Science®

This publication is not indexed in Web of Science®

Scopus

This publication is not indexed in Scopus

Google Scholar

Times Cited: 18

(Last checked: 2026-04-28 08:11)

View record in Google Scholar

This publication is not indexed in Overton

Abstract
We introduce UniversalCEFR, a large-scale multilingual multidimensional dataset of texts annotated according to the CEFR (Common European Framework of Reference) scale in 13 languages. To enable open research in both automated readability and language proficiency assessment, UniversalCEFR comprises 505,807 CEFR-labeled texts curated from educational and learner-oriented resources, standardized into a unified data format to support consistent processing, analysis, and modeling across tasks and languages. To demonstrate its utility, we conduct benchmark experiments using three modelling paradigms: a) linguistic feature-based classification, b) fine-tuning pre-trained LLMs, and c) descriptor-based prompting of instruction-tuned LLMs. Our results further support using linguistic features and fine-tuning pretrained models in multilingual CEFR level assessment. Overall, UniversalCEFR aims to establish best practices in data distribution in language proficiency research by standardising dataset formats and promoting their accessibility to the global research community.
Acknowledgements
--
Keywords
Associated Records

This publication is associated with the following record: