arXiv Open Access 2021

Grammar Accuracy Evaluation (GAE): Quantifiable Quantitative Evaluation of Machine Translation Models

Dojun Park Youngjin Jang Harksoo Kim
Lihat Sumber

Abstrak

Natural Language Generation (NLG) refers to the operation of expressing the calculation results of a system in human language. Since the quality of generated sentences from an NLG model cannot be fully represented using only quantitative evaluation, they are evaluated using qualitative evaluation by humans in which the meaning or grammar of a sentence is scored according to a subjective criterion. Nevertheless, the existing evaluation methods have a problem as a large score deviation occurs depending on the criteria of evaluators. In this paper, we propose Grammar Accuracy Evaluation (GAE) that can provide the specific evaluating criteria. As a result of analyzing the quality of machine translation by BLEU and GAE, it was confirmed that the BLEU score does not represent the absolute performance of machine translation models and GAE compensates for the shortcomings of BLEU with flexible evaluation of alternative synonyms and changes in sentence structure.

Topik & Kata Kunci

Penulis (3)

D

Dojun Park

Y

Youngjin Jang

H

Harksoo Kim

Format Sitasi

Park, D., Jang, Y., Kim, H. (2021). Grammar Accuracy Evaluation (GAE): Quantifiable Quantitative Evaluation of Machine Translation Models. https://arxiv.org/abs/2105.14277

Akses Cepat

Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2021
Bahasa
en
Sumber Database
arXiv
Akses
Open Access ✓