Semantic Scholar Open Access 2015 1953 sitasi

Deep learning and the information bottleneck principle

Naftali Tishby Noga Zaslavsky

Abstrak

Deep Neural Networks (DNNs) are analyzed via the theoretical framework of the information bottleneck (IB) principle. We first show that any DNN can be quantified by the mutual information between the layers and the input and output variables. Using this representation we can calculate the optimal information theoretic limits of the DNN and obtain finite sample generalization bounds. The advantage of getting closer to the theoretical limit is quantifiable both by the generalization bound and by the network's simplicity. We argue that both the optimal architecture, number of layers and features/connections at each layer, are related to the bifurcation points of the information bottleneck tradeoff, namely, relevant compression of the input layer with respect to the output layer. The hierarchical representations at the layered network naturally correspond to the structural phase transitions along the information curve. We believe that this new insight can lead to new optimality bounds and deep learning algorithms.

Penulis (2)

N

Naftali Tishby

N

Noga Zaslavsky

Format Sitasi

Tishby, N., Zaslavsky, N. (2015). Deep learning and the information bottleneck principle. https://doi.org/10.1109/ITW.2015.7133169

Akses Cepat

Lihat di Sumber doi.org/10.1109/ITW.2015.7133169
Informasi Jurnal
Tahun Terbit
2015
Bahasa
en
Total Sitasi
1953×
Sumber Database
Semantic Scholar
DOI
10.1109/ITW.2015.7133169
Akses
Open Access ✓