arXiv Open Access 2025

Learning Decomposed Contextual Token Representations from Pretrained and Collaborative Signals for Generative Recommendation

Yifan Liu Yaokun Liu Zelin Li Zhenrui Yue Gyuseok Lee +3 lainnya
Lihat Sumber

Abstrak

Recent advances in generative recommenders adopt a two-stage paradigm: items are first tokenized into semantic IDs using a pretrained tokenizer, and then large language models (LLMs) are trained to generate the next item via sequence-to-sequence modeling. However, these two stages are optimized for different objectives: semantic reconstruction during tokenizer pretraining versus user interaction modeling during recommender training. This objective misalignment leads to two key limitations: (i) suboptimal static tokenization, where fixed token assignments fail to reflect diverse usage contexts; and (ii) discarded pretrained semantics, where pretrained knowledge - typically from language model embeddings - is overwritten during recommender training on user interactions. To address these limitations, we propose to learn DEcomposed COntextual Token Representations (DECOR), a unified framework that preserves pretrained semantics while enhancing the adaptability of token embeddings. DECOR introduces contextualized token composition to refine token embeddings based on user interaction context, and decomposed embedding fusion that integrates pretrained codebook embeddings with newly learned collaborative embeddings. Experiments on three real-world datasets demonstrate that DECOR consistently outperforms state-of-the-art baselines in recommendation performance. Our code will be made available upon publication.

Topik & Kata Kunci

Penulis (8)

Y

Yifan Liu

Y

Yaokun Liu

Z

Zelin Li

Z

Zhenrui Yue

G

Gyuseok Lee

R

Ruichen Yao

Y

Yang Zhang

D

Dong Wang

Format Sitasi

Liu, Y., Liu, Y., Li, Z., Yue, Z., Lee, G., Yao, R. et al. (2025). Learning Decomposed Contextual Token Representations from Pretrained and Collaborative Signals for Generative Recommendation. https://arxiv.org/abs/2509.10468

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2025
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓