arXiv Open Access 2025

Low-Resource Dialect Adaptation of Large Language Models: A French Dialect Case-Study

Eeham Khan Firas Saidani Owen Van Esbroeck Richard Khoury Leila Kosseim
Lihat Sumber

Abstrak

Despite the widespread adoption of Large Language Models (LLMs), their strongest capabilities remain largely confined to a small number of high-resource languages for which there is abundant training data. Recently, continual pre-training (CPT) has emerged as a means to fine-tune these models to low-resource regional dialects. In this paper, we study the use of CPT for dialect learning under tight data and compute budgets. Using low-rank adaptation (LoRA) and compute-efficient continual pre-training, we adapt three LLMs to the Québec French dialect using a very small dataset and benchmark them on the COLE suite. Our experiments demonstrate an improvement on the minority dialect benchmarks with minimal regression on the prestige language benchmarks with around 1% of model parameters updated. Analysis of the results demonstrate that gains are highly contingent on corpus composition. These findings indicate that CPT with parameter-efficient fine-tuning (PEFT) can narrow the dialect gap by providing cost-effective and sustainable language resource creation, expanding high-quality LLM access to minority linguistic communities. To support reproducibility and broaden access, we release the first Québec French LLMs on Hugging Face.

Topik & Kata Kunci

Penulis (5)

E

Eeham Khan

F

Firas Saidani

O

Owen Van Esbroeck

R

Richard Khoury

L

Leila Kosseim

Format Sitasi

Khan, E., Saidani, F., Esbroeck, O.V., Khoury, R., Kosseim, L. (2025). Low-Resource Dialect Adaptation of Large Language Models: A French Dialect Case-Study. https://arxiv.org/abs/2510.22747

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓