arXiv Open Access 2025

Better Aligned with Survey Respondents or Training Data? Unveiling Political Leanings of LLMs on U.S. Supreme Court Cases

Shanshan Xu T. Y. S. S Santosh Yanai Elazar Quirin Vogel Barbara Plank +1 lainnya
Lihat Sumber

Abstrak

Recent works have shown that Large Language Models (LLMs) have a tendency to memorize patterns and biases present in their training data, raising important questions about how such memorized content influences model behavior. One such concern is the emergence of political bias in LLM outputs. In this paper, we investigate the extent to which LLMs' political leanings reflect memorized patterns from their pretraining corpora. We propose a method to quantitatively evaluate political leanings embedded in the large pretraining corpora. Subsequently we investigate to whom are the LLMs' political leanings more aligned with, their pretrainig corpora or the surveyed human opinions. As a case study, we focus on probing the political leanings of LLMs in 32 US Supreme Court cases, addressing contentious topics such as abortion and voting rights. Our findings reveal that LLMs strongly reflect the political leanings in their training data, and no strong correlation is observed with their alignment to human opinions as expressed in surveys. These results underscore the importance of responsible curation of training data, and the methodology for auditing the memorization in LLMs to ensure human-AI alignment.

Topik & Kata Kunci

Penulis (6)

S

Shanshan Xu

T

T. Y. S. S Santosh

Y

Yanai Elazar

Q

Quirin Vogel

B

Barbara Plank

M

Matthias Grabmair

Format Sitasi

Xu, S., Santosh, T.Y.S.S., Elazar, Y., Vogel, Q., Plank, B., Grabmair, M. (2025). Better Aligned with Survey Respondents or Training Data? Unveiling Political Leanings of LLMs on U.S. Supreme Court Cases. https://arxiv.org/abs/2502.18282

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓