arXiv Open Access 2025

Domain Generalizable Portrait Style Transfer

Xinbo Wang Wenju Xu Qing Zhang Wei-Shi Zheng
Lihat Sumber

Abstrak

This paper presents a portrait style transfer method that generalizes well to various different domains while enabling high-quality semantic-aligned stylization on regions including hair, eyes, eyelashes, skins, lips, and background. To this end, we propose to establish dense semantic correspondence between the given input and reference portraits based on a pre-trained model and a semantic adapter, with which we obtain a warped reference semantically aligned with the input. To ensure effective yet controllable style transfer, we devise an AdaIN-Wavelet transform to balance content preservation and stylization by blending low-frequency information of the warped reference with high-frequency information of the input in the latent space. A style adapter is also designed to provide style guidance from the warped reference. With the stylized latent from AdaIN-Wavelet transform, we employ a dual-conditional diffusion model that integrates a ControlNet recording high-frequency information and the style guidance to generate the final result. Extensive experiments demonstrate the superiority of our method. Our code and trained model are available at https://github.com/wangxb29/DGPST.

Topik & Kata Kunci

Penulis (4)

X

Xinbo Wang

W

Wenju Xu

Q

Qing Zhang

W

Wei-Shi Zheng

Format Sitasi

Wang, X., Xu, W., Zhang, Q., Zheng, W. (2025). Domain Generalizable Portrait Style Transfer. https://arxiv.org/abs/2507.04243

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓