arXiv Open Access 2026

StyleDecoupler: Generalizable Artistic Style Disentanglement

Zexi Jia Jinchao Zhang Jie Zhou
Lihat Sumber

Abstrak

Representing artistic style is challenging due to its deep entanglement with semantic content. We propose StyleDecoupler, an information-theoretic framework that leverages a key insight: multi-modal vision models encode both style and content, while uni-modal models suppress style to focus on content-invariant features. By using uni-modal representations as content-only references, we isolate pure style features from multi-modal embeddings through mutual information minimization. StyleDecoupler operates as a plug-and-play module on frozen Vision-Language Models without fine-tuning. We also introduce WeART, a large-scale benchmark of 280K artworks across 152 styles and 1,556 artists. Experiments show state-of-the-art performance on style retrieval across WeART and WikiART, while enabling applications like style relationship mapping and generative model evaluation. We release our method and dataset at this url.

Topik & Kata Kunci

Penulis (3)

Z

Zexi Jia

J

Jinchao Zhang

J

Jie Zhou

Format Sitasi

Jia, Z., Zhang, J., Zhou, J. (2026). StyleDecoupler: Generalizable Artistic Style Disentanglement. https://arxiv.org/abs/2601.17697

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2026
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓