DOAJ Open Access 2020

Limitations to Estimating Mutual Information in Large Neural Populations

Jan Mölter Geoffrey J. Goodhill

Abstrak

Information theory provides a powerful framework to analyse the representation of sensory stimuli in neural population activity. However, estimating the quantities involved such as entropy and mutual information from finite samples is notoriously hard and any direct estimate is known to be heavily biased. This is especially true when considering large neural populations. We study a simple model of sensory processing and show through a combinatorial argument that, with high probability, for large neural populations any finite number of samples of neural activity in response to a set of stimuli is mutually distinct. As a consequence, the mutual information when estimated directly from empirical histograms will be equal to the stimulus entropy. Importantly, this is the case irrespective of the precise relation between stimulus and neural activity and corresponds to a maximal bias. This argument is general and applies to any application of information theory, where the state space is large and one relies on empirical histograms. Overall, this work highlights the need for alternative approaches for an information theoretic analysis when dealing with large neural populations.

Penulis (2)

J

Jan Mölter

G

Geoffrey J. Goodhill

Format Sitasi

Mölter, J., Goodhill, G.J. (2020). Limitations to Estimating Mutual Information in Large Neural Populations. https://doi.org/10.3390/e22040490

Akses Cepat

PDF tidak tersedia langsung

Cek di sumber asli →
Lihat di Sumber doi.org/10.3390/e22040490
Informasi Jurnal
Tahun Terbit
2020
Sumber Database
DOAJ
DOI
10.3390/e22040490
Akses
Open Access ✓