arXiv Open Access 2025

Natural Language-based Assessment of L2 Oral Proficiency using LLMs

Stefano Bannò Rao Ma Mengjie Qian Siyuan Tang Kate Knill +1 lainnya
Lihat Sumber

Abstrak

Natural language-based assessment (NLA) is an approach to second language assessment that uses instructions - expressed in the form of can-do descriptors - originally intended for human examiners, aiming to determine whether large language models (LLMs) can interpret and apply them in ways comparable to human assessment. In this work, we explore the use of such descriptors with an open-source LLM, Qwen 2.5 72B, to assess responses from the publicly available S&I Corpus in a zero-shot setting. Our results show that this approach - relying solely on textual information - achieves competitive performance: while it does not outperform state-of-the-art speech LLMs fine-tuned for the task, it surpasses a BERT-based model trained specifically for this purpose. NLA proves particularly effective in mismatched task settings, is generalisable to other data types and languages, and offers greater interpretability, as it is grounded in clearly explainable, widely applicable language descriptors.

Topik & Kata Kunci

Penulis (6)

S

Stefano Bannò

R

Rao Ma

M

Mengjie Qian

S

Siyuan Tang

K

Kate Knill

M

Mark Gales

Format Sitasi

Bannò, S., Ma, R., Qian, M., Tang, S., Knill, K., Gales, M. (2025). Natural Language-based Assessment of L2 Oral Proficiency using LLMs. https://arxiv.org/abs/2507.10200

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓