arXiv Open Access 2022

Non-Linguistic Supervision for Contrastive Learning of Sentence Embeddings

Yiren Jian Chongyang Gao Soroush Vosoughi
Lihat Sumber

Abstrak

Semantic representation learning for sentences is an important and well-studied problem in NLP. The current trend for this task involves training a Transformer-based sentence encoder through a contrastive objective with text, i.e., clustering sentences with semantically similar meanings and scattering others. In this work, we find the performance of Transformer models as sentence encoders can be improved by training with multi-modal multi-task losses, using unpaired examples from another modality (e.g., sentences and unrelated image/audio data). In particular, besides learning by the contrastive loss on text, our model clusters examples from a non-linguistic domain (e.g., visual/audio) with a similar contrastive loss at the same time. The reliance of our framework on unpaired non-linguistic data makes it language-agnostic, enabling it to be widely applicable beyond English NLP. Experiments on 7 semantic textual similarity benchmarks reveal that models trained with the additional non-linguistic (images/audio) contrastive objective lead to higher quality sentence embeddings. This indicates that Transformer models are able to generalize better by doing a similar task (i.e., clustering) with unpaired examples from different modalities in a multi-task fashion.

Topik & Kata Kunci

Penulis (3)

Y

Yiren Jian

C

Chongyang Gao

S

Soroush Vosoughi

Format Sitasi

Jian, Y., Gao, C., Vosoughi, S. (2022). Non-Linguistic Supervision for Contrastive Learning of Sentence Embeddings. https://arxiv.org/abs/2209.09433

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2022
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓