Semantic Scholar Open Access 2016 987 sitasi

Siamese Recurrent Architectures for Learning Sentence Similarity

Jonas W. Mueller Aditya Thyagarajan

Abstrak

We present a siamese adaptation of the Long Short-Term Memory (LSTM) network for labeled data comprised of pairs of variable-length sequences. Our model is applied to assess semantic similarity between sentences, where we exceed state of the art, outperforming carefully handcrafted features and recently proposed neural network systems of greater complexity. For these applications, we provide word-embedding vectors supplemented with synonymic information to the LSTMs, which use a fixed size vector to encode the underlying meaning expressed in a sentence (irrespective of the particular wording/syntax). By restricting subsequent operations to rely on a simple Manhattan metric, we compel the sentence representations learned by our model to form a highly structured space whose geometry reflects complex semantic relationships. Our results are the latest in a line of findings that showcase LSTMs as powerful language models capable of tasks requiring intricate understanding.

Topik & Kata Kunci

Penulis (2)

J

Jonas W. Mueller

A

Aditya Thyagarajan

Format Sitasi

Mueller, J.W., Thyagarajan, A. (2016). Siamese Recurrent Architectures for Learning Sentence Similarity. https://doi.org/10.1609/aaai.v30i1.10350

Akses Cepat

Lihat di Sumber doi.org/10.1609/aaai.v30i1.10350
Informasi Jurnal
Tahun Terbit
2016
Bahasa
en
Total Sitasi
987×
Sumber Database
Semantic Scholar
DOI
10.1609/aaai.v30i1.10350
Akses
Open Access ✓