arXiv Open Access 2025

Emissions and Performance Trade-off Between Small and Large Language Models

Anandita Garg Uma Gaba Deepan Muthirayan Anish Roy Chowdhury
Lihat Sumber

Abstrak

The advent of Large Language Models (LLMs) has raised concerns about their enormous carbon footprint, starting with energy-intensive training and continuing through repeated inference. This study investigates the potential of using fine-tuned Small Language Models (SLMs) as a sustainable alternative for predefined tasks. Here, we present a comparative analysis of the performance-emissions trade-off between LLMs and fine-tuned SLMs across selected tasks under Natural Language Processing, Reasoning and Programming. Our results show that in four out of the six selected tasks, SLMs maintained comparable performances for a significant reduction in carbon emissions during inference. Our findings demonstrate the viability of smaller models in mitigating the environmental impact of resource-heavy LLMs, thus advancing towards sustainable, green AI.

Penulis (4)

A

Anandita Garg

U

Uma Gaba

D

Deepan Muthirayan

A

Anish Roy Chowdhury

Format Sitasi

Garg, A., Gaba, U., Muthirayan, D., Chowdhury, A.R. (2025). Emissions and Performance Trade-off Between Small and Large Language Models. https://arxiv.org/abs/2601.08844

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓