arXiv Open Access 2023

Discourse Structure Extraction from Pre-Trained and Fine-Tuned Language Models in Dialogues

Chuyuan Li Patrick Huber Wen Xiao Maxime Amblard Chloé Braud +1 lainnya
Lihat Sumber

Abstrak

Discourse processing suffers from data sparsity, especially for dialogues. As a result, we explore approaches to build discourse structures for dialogues, based on attention matrices from Pre-trained Language Models (PLMs). We investigate multiple tasks for fine-tuning and show that the dialogue-tailored Sentence Ordering task performs best. To locate and exploit discourse information in PLMs, we propose an unsupervised and a semi-supervised method. Our proposals achieve encouraging results on the STAC corpus, with F1 scores of 57.2 and 59.3 for unsupervised and semi-supervised methods, respectively. When restricted to projective trees, our scores improved to 63.3 and 68.1.

Topik & Kata Kunci

Penulis (6)

C

Chuyuan Li

P

Patrick Huber

W

Wen Xiao

M

Maxime Amblard

C

Chloé Braud

G

Giuseppe Carenini

Format Sitasi

Li, C., Huber, P., Xiao, W., Amblard, M., Braud, C., Carenini, G. (2023). Discourse Structure Extraction from Pre-Trained and Fine-Tuned Language Models in Dialogues. https://arxiv.org/abs/2302.05895

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2023
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓