arXiv Open Access 2026

LLMs Can Infer Political Alignment from Online Conversations

Byunghwee Lee Sangyeon Kim Filippo Menczer Yong-Yeol Ahn Haewoon Kwak +1 lainnya
Lihat Sumber

Abstrak

Due to the correlational structure in our traits such as identities, cultures, and political attitudes, seemingly innocuous preferences like following a band or using a specific slang can reveal private traits. This possibility, especially when combined with massive, public social data and advanced computational methods, poses a fundamental privacy risk. As our data exposure online and the rapid advancement of AI are increasing the risk of misuse, it is critical to understand the capacity of large language models (LLMs) to exploit such potential. Here, using online discussions on DebateOrg and Reddit, we show that LLMs can reliably infer hidden political alignment, significantly outperforming traditional machine learning models. Prediction accuracy further improves as we aggregate multiple text-level inferences into a user-level prediction, and as we use more politics-adjacent domains. We demonstrate that LLMs leverage words that are highly predictive of political alignment while not being explicitly political. Our findings underscore the capacity and risks of LLMs for exploiting socio-cultural correlates.

Topik & Kata Kunci

Penulis (6)

B

Byunghwee Lee

S

Sangyeon Kim

F

Filippo Menczer

Y

Yong-Yeol Ahn

H

Haewoon Kwak

J

Jisun An

Format Sitasi

Lee, B., Kim, S., Menczer, F., Ahn, Y., Kwak, H., An, J. (2026). LLMs Can Infer Political Alignment from Online Conversations. https://arxiv.org/abs/2603.11253

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2026
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓