arXiv Open Access 2024

Inference-Time Language Model Alignment via Integrated Value Guidance

Zhixuan Liu Zhanhui Zhou Yuanfu Wang Chao Yang Yu Qiao
Lihat Sumber

Abstrak

Large language models are typically fine-tuned to align with human preferences, but tuning large models is computationally intensive and complex. In this work, we introduce $\textit{Integrated Value Guidance}$ (IVG), a method that uses implicit and explicit value functions to guide language model decoding at token and chunk-level respectively, efficiently aligning large language models purely at inference time. This approach circumvents the complexities of direct fine-tuning and outperforms traditional methods. Empirically, we demonstrate the versatility of IVG across various tasks. In controlled sentiment generation and summarization tasks, our method significantly improves the alignment of large models using inference-time guidance from $\texttt{gpt2}$-based value functions. Moreover, in a more challenging instruction-following benchmark AlpacaEval 2.0, we show that both specifically tuned and off-the-shelf value functions greatly improve the length-controlled win rates of large models against $\texttt{gpt-4-turbo}$ (e.g., $19.51\% \rightarrow 26.51\%$ for $\texttt{Mistral-7B-Instruct-v0.2}$ and $25.58\% \rightarrow 33.75\%$ for $\texttt{Mixtral-8x7B-Instruct-v0.1}$ with Tulu guidance).

Topik & Kata Kunci

Penulis (5)

Z

Zhixuan Liu

Z

Zhanhui Zhou

Y

Yuanfu Wang

C

Chao Yang

Y

Yu Qiao

Format Sitasi

Liu, Z., Zhou, Z., Wang, Y., Yang, C., Qiao, Y. (2024). Inference-Time Language Model Alignment via Integrated Value Guidance. https://arxiv.org/abs/2409.17819

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2024
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓