Semantic Scholar Open Access 2020 7571 sitasi

Measuring Massive Multitask Language Understanding

Dan Hendrycks Collin Burns Steven Basart Andy Zou Mantas Mazeika +2 lainnya

Abstrak

We propose a new test to measure a text model's multitask accuracy. The test covers 57 tasks including elementary mathematics, US history, computer science, law, and more. To attain high accuracy on this test, models must possess extensive world knowledge and problem solving ability. We find that while most recent models have near random-chance accuracy, the very largest GPT-3 model improves over random chance by almost 20 percentage points on average. However, on every one of the 57 tasks, the best models still need substantial improvements before they can reach expert-level accuracy. Models also have lopsided performance and frequently do not know when they are wrong. Worse, they still have near-random accuracy on some socially important subjects such as morality and law. By comprehensively evaluating the breadth and depth of a model's academic and professional understanding, our test can be used to analyze models across many tasks and to identify important shortcomings.

Topik & Kata Kunci

Penulis (7)

D

Dan Hendrycks

C

Collin Burns

S

Steven Basart

A

Andy Zou

M

Mantas Mazeika

D

D. Song

J

J. Steinhardt

Format Sitasi

Hendrycks, D., Burns, C., Basart, S., Zou, A., Mazeika, M., Song, D. et al. (2020). Measuring Massive Multitask Language Understanding. https://www.semanticscholar.org/paper/814a4f680b9ba6baba23b93499f4b48af1a27678

Akses Cepat

PDF tidak tersedia langsung

Cek di sumber asli →
Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2020
Bahasa
en
Total Sitasi
7571×
Sumber Database
Semantic Scholar
Akses
Open Access ✓