Semantic Scholar Open Access 2017 7566 sitasi

On Calibration of Modern Neural Networks

Chuan Guo Geoff Pleiss Yu Sun Kilian Q. Weinberger

Abstrak

Confidence calibration -- the problem of predicting probability estimates representative of the true correctness likelihood -- is important for classification models in many applications. We discover that modern neural networks, unlike those from a decade ago, are poorly calibrated. Through extensive experiments, we observe that depth, width, weight decay, and Batch Normalization are important factors influencing calibration. We evaluate the performance of various post-processing calibration methods on state-of-the-art architectures with image and document classification datasets. Our analysis and experiments not only offer insights into neural network learning, but also provide a simple and straightforward recipe for practical settings: on most datasets, temperature scaling -- a single-parameter variant of Platt Scaling -- is surprisingly effective at calibrating predictions.

Penulis (4)

C

Chuan Guo

G

Geoff Pleiss

Y

Yu Sun

K

Kilian Q. Weinberger

Format Sitasi

Guo, C., Pleiss, G., Sun, Y., Weinberger, K.Q. (2017). On Calibration of Modern Neural Networks. https://www.semanticscholar.org/paper/d65ce2b8300541414bfe51d03906fca72e93523c

Akses Cepat

PDF tidak tersedia langsung

Cek di sumber asli →
Lihat di Sumber
Informasi Jurnal
Tahun Terbit
2017
Bahasa
en
Total Sitasi
7566×
Sumber Database
Semantic Scholar
Akses
Open Access ✓