Estimation of information-theoretic quantities
Liam Paninski
Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/∼liam liam@gatsby.ucl.ac.uk November 16, 2004
Estimation of information-theoretic quantities Liam Paninski - - PowerPoint PPT Presentation
Estimation of information-theoretic quantities Liam Paninski Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/ liam liam@gatsby.ucl.ac.uk November 16, 2004 Estimation of information Some
Gatsby Computational Neuroscience Unit University College London http://www.gatsby.ucl.ac.uk/∼liam liam@gatsby.ucl.ac.uk November 16, 2004
(Warland et al., 1997)
m
0.2 0.4 0.6 0.8 Sample distributions of MLE; p uniform; m=500 N=10 0.05 0.1 P(Hest) N=100 0.05 0.1 0.15 N=500 1 2 3 4 5 6 7 8 0.05 0.1 0.15 Hest (bits) N=1000
N = number of samples
0.5 1 1 2 3 4 5 m=N=100 20 40 60 80 100 1 2 3 4 0.5 1 2 4 6 m=N=1000 200 400 600 800 1000 1 2 3 4 5 n 0.5 1 2 4 6 sorted, normalized p m=N=10000 2000 4000 6000 8000 10000 2 4 6 unsorted, unnormalized i
N log t N .
j
j g(j)Bj(p) close to f(p) for all p, bias will be small
i g(ni);
i g(ni)) small
j g(j)Bj(t)|
10
1
10
2
10
3
10
4
10
5
10
6
0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2 N RMS error bound (bits) Upper and lower bounds on maximum rms error; N/m = 0.25 BUB JK
10
1
10
2
10
3
10
4
10
5
10
6
0.2 0.4 0.6 0.8 1 1.2 1.4 1.6 1.8 2 2.2 N RMS error bound (bits) Upper (BUB) and lower (JK) bounds on maximum rms error N/m = 0.10 (BUB) N/m = 0.25 (BUB) N/m = 1.00 (BUB) N/m = 0.10 (JK) N/m = 0.25 (JK) N/m = 1.00 (JK)
10
2
3 3.5 4 4.5 5 5.5 6 6.5 True entropy bits 10
2
−3 −2.5 −2 −1.5 −1 −0.5 Bias 10
2
0.05 0.1 0.15 0.2 0.25 0.3 Standard deviation bits firing rate (Hz) 10
2
0.5 1 1.5 2 2.5 3 RMS error firing rate (Hz) MLE MM JK BUB
y x true p(y | x) 2 4 6 8 10 2 4 6 8 10 y estimated p(y | x) 2 4 6 8 10 y | error | 2 4 6 8 10 0.002 0.004 0.006 0.008 0.01 0.012
mx = my = 700; N/mxy = 0.3 ˆ IMLE = 2.21 bits ˆ IMM = −0.19 bits ˆ IBUB = 0.60 bits; conservative (worst-case upper bound) error: ±0.2 bits true I(X; Y ) = 0.62 bits
Beirlant, J., Dudewicz, E., Gyorfi, L., and van der Meulen, E. (1997). Nonparametric entropy estimation: an
Miller, G. (1955). Note on the bias of information estimates. In Information theory in psychology II-B, pages 95–100. Nemenman, I., Shafee, F., and Bialek, W. (2002). Entropy and inference, revisited. Advances in neural information processing, 14. Paninski, L. (2003). Estimation of entropy and mutual information. Neural Computation, 15:1191–1253. Paninski, L. (2004). Estimating entropy on m bins given fewer than m samples. IEEE Transactions on Information Theory, 50:2200–2203. Rieke, F., Warland, D., de Ruyter van Steveninck, R., and Bialek, W. (1997). Spikes: Exploring the neural code. MIT Press, Cambridge. Strong, S. Koberle, R., de Ruyter van Steveninck R., and Bialek, W. (1998). Entropy and information in neural spike trains. Physical Review Letters, 80:197–202. Treves, A. and Panzeri, S. (1995). The upward bias in measures of information derived from limited data
Warland, D., Reinagel, P., and Meister, M. (1997). Decoding visual information from a population of retinal ganglion cells. Journal of Neurophysiology, 78:2336–2350. Wolpert, D. and Wolf, D. (1995). Estimating functions of probability distributions from a finite set of