arXiv Open Access 2023

Enhanced Transformer Architecture for Natural Language Processing

Woohyeon Moon Taeyoung Kim Bumgeun Park Dongsoo Har
Lihat Sumber

Abstrak

Transformer is a state-of-the-art model in the field of natural language processing (NLP). Current NLP models primarily increase the number of transformers to improve processing performance. However, this technique requires a lot of training resources such as computing capacity. In this paper, a novel structure of Transformer is proposed. It is featured by full layer normalization, weighted residual connection, positional encoding exploiting reinforcement learning, and zero masked self-attention. The proposed Transformer model, which is called Enhanced Transformer, is validated by the bilingual evaluation understudy (BLEU) score obtained with the Multi30k translation dataset. As a result, the Enhanced Transformer achieves 202.96% higher BLEU score as compared to the original transformer with the translation dataset.

Topik & Kata Kunci

Penulis (4)

W

Woohyeon Moon

T

Taeyoung Kim

B

Bumgeun Park

D

Dongsoo Har

Format Sitasi

Moon, W., Kim, T., Park, B., Har, D. (2023). Enhanced Transformer Architecture for Natural Language Processing. https://arxiv.org/abs/2310.10930

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2023
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓