arXiv Open Access 2025

Towards Fundamental Language Models: Does Linguistic Competence Scale with Model Size?

Jaime Collado-Montañez L. Alfonso Ureña-López Arturo Montejo-Ráez
Lihat Sumber

Abstrak

Large Language Models offer impressive language capabilities but suffer from well-known limitations, including hallucinations, biases, privacy concerns, and high computational costs. These issues are largely driven by the combination of linguistic competence and factual memorization within a single monolithic model. This paper introduces and empirically supports the Fundamental Language Model (FLM) paradigm, which advocates for smaller, linguistically competent models that offload factual retrieval to external tools. We evaluate models ranging from 135M to 32B parameters across three dimensions: linguistic competence, external factual knowledge, and internal factual knowledge. Our findings reveal that while both linguistic competence and factual knowledge improve with scale, internal factual knowledge grows significantly faster, suggesting that model size is more closely tied to memorization than to core language ability. These results support a modular approach to language modeling, where compact, linguistically proficient models serve as the foundation for tool-augmented systems. The FLM paradigm offers a path toward more efficient, interpretable, and sustainable NLP solutions.

Topik & Kata Kunci

Penulis (3)

J

Jaime Collado-Montañez

L

L. Alfonso Ureña-López

A

Arturo Montejo-Ráez

Format Sitasi

Collado-Montañez, J., Ureña-López, L.A., Montejo-Ráez, A. (2025). Towards Fundamental Language Models: Does Linguistic Competence Scale with Model Size?. https://arxiv.org/abs/2509.02225

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓