Neural encoding models & maximum likelihood Jonathan Pillow 1 - - PowerPoint PPT Presentation

neural encoding models maximum likelihood
SMART_READER_LITE
LIVE PREVIEW

Neural encoding models & maximum likelihood Jonathan Pillow 1 - - PowerPoint PPT Presentation

Statistical modeling and analysis of neural data NEU 560, Spring 2018 Lecture 7 Neural encoding models & maximum likelihood Jonathan Pillow 1 probability leftovers: sampling vs inference Model Data 700 samples sampling (measurement)


slide-1
SLIDE 1

Neural encoding models & maximum likelihood

Statistical modeling and analysis of neural data NEU 560, Spring 2018 Lecture 7 Jonathan Pillow

1

slide-2
SLIDE 2

true mean: [0 0.8] true cov: [1.0 -0.25

  • 0.25 0.3]

sample mean: [-0.05 0.83] sample cov: [0.95 -0.23

  • 0.23 0.29]

700 samples

sampling (measurement) Inference (“fitting”)

probability leftovers: sampling vs inference

Model Data

2

slide-3
SLIDE 3

Estimation

( 1

2

, , r r = r



neuron # spike count

parameter

(“stimulus”)

measured dataset


(“population response”)

An estimator is a function

model

  • often we will write or just

ˆ θ(x)

p(x|θ) x = {r1, r2, . . . , rn}

f : x − → ˆ θ

3

slide-4
SLIDE 4

Properties of an estimator

bias:

  • “unbiased” if bias=0

variance:

  • “consistent” if bias and variance both go


to zero asymptotically

Q: what is the bias of the estimator 
 (i.e., estimate is 7 for all datasets x)

ˆ θ(x) = 7

“expected” value 
 (average over draws of x)

Q: what is the variance of that estimator?

4

slide-5
SLIDE 5

y x

stimuli spike trains

neural coding problem

Q: what is the probabilistic relationship between stimuli and spike trains?

5

slide-6
SLIDE 6

y x

stimuli spike trains

neural coding problem

Q: what is the probabilistic relationship between stimuli and spike trains?

“encoding model”

6

slide-7
SLIDE 7

y x

stimuli spike trains

today: single-neuron encoding

“encoding model”

P(yi|~ x, ✓)

Question: what criteria for picking a model?

7

slide-8
SLIDE 8

model desiderata

fittability / tractability richness / flexibility

(can be fit to data) (capture realistic neural properties)

linear, Gaussian multi- compartment Hodgkin-Huxley sweet spot

GLM

8

slide-9
SLIDE 9

Example 1: linear Poisson neuron

spike count spike rate encoding model: stimulus parameter

−3 −2 − 1 2 3 4 5 6 7 8 9 10

= mean = variance

9

slide-10
SLIDE 10

20 40 20 40 60

(contrast) (spike count)

20 40 60

conditional distribution

p(y|x)

10

slide-11
SLIDE 11

20 40 20 40 60

(contrast) (spike count)

20 40 60

conditional distribution

p(y|x)

11

slide-12
SLIDE 12

20 40 20 40 60

(contrast) (spike count)

20 40 60

conditional distribution

p(y|x)

12

slide-13
SLIDE 13

Maximum Likelihood Estimation:

  • given observed data , find that maximizes

all spike counts all stimuli parameters

}

single-trial probability

Q: what assumption are we making about the responses? A: conditional independence across trials!

13

slide-14
SLIDE 14

Q: when do we call a likelihood? Maximum Likelihood Estimation:

  • given observed data , find that maximizes

all spike counts all stimuli parameters

}

single-trial probability

Q: what assumption are we making about the responses? A: conditional independence across trials! A: when considering it as a function of !

14

slide-15
SLIDE 15

20 40 20 40 60

(contrast) (spike count)

Maximum Likelihood Estimation:

  • given observed data , find that maximizes

p(y|x)

  • could in theory do this by turning a knob

15

slide-16
SLIDE 16

20 40 20 40 60

(contrast) (spike count)

Maximum Likelihood Estimation:

  • given observed data , find that maximizes

p(y|x)

  • could in theory do this by turning a knob

16

slide-17
SLIDE 17

20 40 20 40 60

(contrast) (spike count)

Maximum Likelihood Estimation:

  • given observed data , find that maximizes

p(y|x)

  • could in theory do this by turning a knob

17

slide-18
SLIDE 18

likelihood

Likelihood function: as a function of

Because data are independent:

1 2

18

slide-19
SLIDE 19

1 2

log-likelihood log

Likelihood function: as a function of

Because data are independent:

1 2

likelihood

19

slide-20
SLIDE 20

1 2

log-likelihood

Do it: solve for

20

slide-21
SLIDE 21
  • Closed-form solution when model in “exponential family”

1 2

log-likelihood

21

slide-22
SLIDE 22

Properties of the MLE (maximum likelihood estimator)

  • consistent

(converges to true in limit of infinite data)

  • efficient 


(converges as quickly as possible, 
 i.e., achieves minimum possible asymptotic error)

22

slide-23
SLIDE 23

Example 2: linear Gaussian neuron

spike count spike rate encoding model: stimulus parameter

23

slide-24
SLIDE 24

20 40 20 40 60

(contrast) (spike count)

20 40 60

All slices have same width encoding distribution

24

slide-25
SLIDE 25

Log-Likelihood Do it: differentiate, set to zero, and solve.

25

slide-26
SLIDE 26

Log-Likelihood Maximum-Likelihood Estimator:

(“Least squares regression” solution) (Recall that for Poisson, )

26

slide-27
SLIDE 27

Example 3: unknown neuron

  • 25

25 25 50 75 100 (contrast) (spike count)

Be the computational neuroscientist: what model would you use?

27

slide-28
SLIDE 28

Example 3: unknown neuron

More general setup:

  • 25

25 25 50 75 100 (contrast) (spike count)

for some nonlinear function f

28

slide-29
SLIDE 29

Answer: stimulus likelihood function - useful for ML stimulus decoding!

Quick Quiz:

  • 1. as a function of y?
  • 2. as a function of ?

The distribution P(y|x, ) can be considered as a function of y, x, or .

spikes stimulus parameters

Answer: encoding distribution - probability distribution over spike counts Answer: likelihood function - the probability of the data given model params

  • 3. as a function of x?

What is P(y|x, ) :

29

slide-30
SLIDE 30

20 40 20 40 60

(contrast) (spike count)

stimulus decoding likelihood

30