arXiv Open Access 2024

Continuous Speculative Decoding for Autoregressive Image Generation

Zili Wang Robert Zhang Kun Ding Qi Yang Fei Li +1 lainnya
Lihat Sumber

Abstrak

Continuous visual autoregressive (AR) models have demonstrated promising performance in image generation. However, the heavy autoregressive inference burden imposes significant overhead. In Large Language Models (LLMs), speculative decoding has effectively accelerated discrete autoregressive inference. However, the absence of an analogous theory for continuous distributions precludes its use in accelerating continuous AR models. To fill this gap, this work presents continuous speculative decoding, and addresses challenges from: 1) low acceptance rate, caused by inconsistent output distribution between target and draft models, and 2) modified distribution without analytic expression, caused by complex integral. To address challenge 1), we propose denoising trajectory alignment and token pre-filling strategies. To address challenge 2), we introduce acceptance-rejection sampling algorithm with an appropriate upper bound, thereby avoiding explicitly calculating the integral. Furthermore, our denoising trajectory alignment is also reused in acceptance-rejection sampling, effectively avoiding repetitive diffusion model inference. Extensive experiments demonstrate that our proposed continuous speculative decoding achieves over $2\times$ speedup on off-the-shelf models, while maintaining the original generation quality. Codes is available at: https://github.com/MarkXCloud/CSpD

Topik & Kata Kunci

Penulis (6)

Z

Zili Wang

R

Robert Zhang

K

Kun Ding

Q

Qi Yang

F

Fei Li

S

Shiming Xiang

Format Sitasi

Wang, Z., Zhang, R., Ding, K., Yang, Q., Li, F., Xiang, S. (2024). Continuous Speculative Decoding for Autoregressive Image Generation. https://arxiv.org/abs/2411.11925

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2024
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓