arXiv Open Access 2026

Rethinking Preference Alignment for Diffusion Models with Classifier-Free Guidance

Zhou Jiang Yandong Wen Zhen Liu
Lihat Sumber

Abstrak

Aligning large-scale text-to-image diffusion models with nuanced human preferences remains challenging. While direct preference optimization (DPO) is simple and effective, large-scale finetuning often shows a generalization gap. We take inspiration from test-time guidance and cast preference alignment as classifier-free guidance (CFG): a finetuned preference model acts as an external control signal during sampling. Building on this view, we propose a simple method that improves alignment without retraining the base model. To further enhance generalization, we decouple preference learning into two modules trained on positive and negative data, respectively, and form a \emph{contrastive guidance} vector at inference by subtracting their predictions (positive minus negative), scaled by a user-chosen strength and added to the base prediction at each step. This yields a sharper and controllable alignment signal. We evaluate on Stable Diffusion 1.5 and Stable Diffusion XL with Pick-a-Pic v2 and HPDv3, showing consistent quantitative and qualitative gains.

Topik & Kata Kunci

Penulis (3)

Z

Zhou Jiang

Y

Yandong Wen

Z

Zhen Liu

Format Sitasi

Jiang, Z., Wen, Y., Liu, Z. (2026). Rethinking Preference Alignment for Diffusion Models with Classifier-Free Guidance. https://arxiv.org/abs/2602.18799

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2026
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓