arXiv Open Access 2025

Emergent Response Planning in LLMs

Zhichen Dong Zhanhui Zhou Zhixuan Liu Chao Yang Chaochao Lu
Lihat Sumber

Abstrak

In this work, we argue that large language models (LLMs), though trained to predict only the next token, exhibit emergent planning behaviors: $\textbf{their hidden representations encode future outputs beyond the next token}$. Through simple probing, we demonstrate that LLM prompt representations encode global attributes of their entire responses, including $\textit{structure attributes}$ (e.g., response length, reasoning steps), $\textit{content attributes}$ (e.g., character choices in storywriting, multiple-choice answers at the end of response), and $\textit{behavior attributes}$ (e.g., answer confidence, factual consistency). In addition to identifying response planning, we explore how it scales with model size across tasks and how it evolves during generation. The findings that LLMs plan ahead for the future in their hidden representations suggest potential applications for improving transparency and generation control.

Topik & Kata Kunci

Penulis (5)

Z

Zhichen Dong

Z

Zhanhui Zhou

Z

Zhixuan Liu

C

Chao Yang

C

Chaochao Lu

Format Sitasi

Dong, Z., Zhou, Z., Liu, Z., Yang, C., Lu, C. (2025). Emergent Response Planning in LLMs. https://arxiv.org/abs/2502.06258

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓