Semantic Scholar Open Access 2022 8731 sitasi

Hierarchical Text-Conditional Image Generation with CLIP Latents

A. Ramesh Prafulla Dhariwal Alex Nichol C. Chu Mark Chen

Abstrak

Contrastive models like CLIP have been shown to learn robust representations of images that capture both semantics and style. To leverage these representations for image generation, we propose a two-stage model: a prior that generates a CLIP image embedding given a text caption, and a decoder that generates an image conditioned on the image embedding. We show that explicitly generating image representations improves image diversity with minimal loss in photorealism and caption similarity. Our decoders conditioned on image representations can also produce variations of an image that preserve both its semantics and style, while varying the non-essential details absent from the image representation. Moreover, the joint embedding space of CLIP enables language-guided image manipulations in a zero-shot fashion. We use diffusion models for the decoder and experiment with both autoregressive and diffusion models for the prior, finding that the latter are computationally more efficient and produce higher-quality samples.

Topik & Kata Kunci

Penulis (5)

A

A. Ramesh

P

Prafulla Dhariwal

A

Alex Nichol

C

C. Chu

M

Mark Chen

Format Sitasi

Ramesh, A., Dhariwal, P., Nichol, A., Chu, C., Chen, M. (2022). Hierarchical Text-Conditional Image Generation with CLIP Latents. https://doi.org/10.48550/arXiv.2204.06125

Akses Cepat

Lihat di Sumber doi.org/10.48550/arXiv.2204.06125
Informasi Jurnal
Tahun Terbit
2022
Bahasa
en
Total Sitasi
8731×
Sumber Database
Semantic Scholar
DOI
10.48550/arXiv.2204.06125
Akses
Open Access ✓