DOAJ Open Access 2022

Temporal Effects on Pre-trained Models for Language Processing Tasks

Oshin Agarwal Ani Nenkova

Abstrak

AbstractKeeping the performance of language technologies optimal as time passes is of great practical interest. We study temporal effects on model performance on downstream language tasks, establishing a nuanced terminology for such discussion and identifying factors essential to conduct a robust study. We present experiments for several tasks in English where the label correctness is not dependent on time and demonstrate the importance of distinguishing between temporal model deterioration and temporal domain adaptation for systems using pre-trained representations. We find that, depending on the task, temporal model deterioration is not necessarily a concern. Temporal domain adaptation, however, is beneficial in all cases, with better performance for a given time period possible when the system is trained on temporally more recent data. Therefore, we also examine the efficacy of two approaches for temporal domain adaptation without human annotations on new data. Self-labeling shows consistent improvement and notably, for named entity recognition, leads to better temporal adaptation than even human annotations.

Penulis (2)

O

Oshin Agarwal

A

Ani Nenkova

Format Sitasi

Agarwal, O., Nenkova, A. (2022). Temporal Effects on Pre-trained Models for Language Processing Tasks. https://doi.org/10.1162/tacl_a_00497

Akses Cepat

PDF tidak tersedia langsung

Cek di sumber asli →
Lihat di Sumber doi.org/10.1162/tacl_a_00497
Informasi Jurnal
Tahun Terbit
2022
Sumber Database
DOAJ
DOI
10.1162/tacl_a_00497
Akses
Open Access ✓