Inferring Inference Xaq Pitkow Rajkumar Vasudeva Raju part of the - - PowerPoint PPT Presentation

inferring inference
SMART_READER_LITE
LIVE PREVIEW

Inferring Inference Xaq Pitkow Rajkumar Vasudeva Raju part of the - - PowerPoint PPT Presentation

Inferring Inference Xaq Pitkow Rajkumar Vasudeva Raju part of the MICrONS project with Tolias, Bethge, Patel, Zemel, Urtasun, Xu, Siapas, Paninski, Baraniuk, Reid, Seung NICE workshop 2017 World Brain match Hypothesis: The brain


slide-1
SLIDE 1

Xaq Pitkow Rajkumar Vasudeva Raju

Inferring Inference

NICE workshop 2017

part of the MICrONS project with Tolias, Bethge, Patel, Zemel, Urtasun, Xu, Siapas, Paninski, Baraniuk, Reid, Seung

slide-2
SLIDE 2

Brain World Hypothesis:

The brain approximates probabilistic inference

  • ver a probabilistic graphical model

using a message-passing algorithm implicit in population dynamics

match

slide-3
SLIDE 3

What algorithms can we learn from the brain?

Architectures? cortex, hippocampus, cerebellum, basal ganglia, … Transformations? nonlinear dynamics from population responses Learning rules? short and long-term plasticity

M B S M B

slide-4
SLIDE 4

Principles: Probabilistic Nonlinear Distributed Details: Graphical models Message-passing inference Multiplexed across neurons

slide-5
SLIDE 5

So neural computation is inevitably statistical. This provides us with mathematical predictions.

world brain

Events in the world can cause many neural responses. Neural responses can be caused by many events.

? ?

slide-6
SLIDE 6

Why does it matter whether processing is linear or nonlinear? If all computation were linear we wouldn’t need a brain.

nonlinearly separable

apples

  • ranges

linearly separable

slide-7
SLIDE 7
slide-8
SLIDE 8
slide-9
SLIDE 9
slide-10
SLIDE 10
slide-11
SLIDE 11

Two sources of nonlinearities

Product rule: p(x,y) = p(x) ∙ p(y) Sum rule: L(x) = log ∑y exp L(x,y)

Relationships between uncertainties posteriors generally have nonlinear dependencies even for the simplest variables Relationships between latent variables Image = Light × Reflectance

I L R

slide-12
SLIDE 12

Probabilistic Graphical Models: Simplify joint distribution p(x|r) by specifying how variables interact

p(x|r) ∝ Y

α

ψα(xα)

slide-13
SLIDE 13

x1 x2 x3 ψ123 Variable Factor

slide-14
SLIDE 14

Example: Pairwise Markov Random Field

x1

x2 x3

J1 J2 J3 J12 J23

slide-15
SLIDE 15
  • Localize information so it is actionable
  • Summarize statistics relevant for targets
  • Send that information along graph
  • Iteratively update factors with new information

Approximate inference by message-passing:

message-passing parameters interactions posterior for neighbors general equation posterior parameters

slide-16
SLIDE 16

Example message-passing algorithms

  • Mean-field (assumes variables are independent)
  • Belief propagation (assumes tree graph)
  • Expectation propagation (updates parametric posterior)
  • Brain’s clever tricks?
slide-17
SLIDE 17

ri

a.r b.r b.r µ = a.r

Neuron index i

p(x|r) x

Neural response Posterior

1 σ = a.r

Spatial representation of uncertainty (e.g. Probabilistic Population Codes, PPCs)

Ma, Beck, Latham, Pouget 2006, etc

Pattern of activity represents probability. More spikes generally means more certainty

slide-18
SLIDE 18

Message-passing updates embedding Neural dynamics

slide-19
SLIDE 19

linear connections singleton populations pairwise populations nonlinear connections

r

12

r

23

r

1

r

2

r

3

x1

x2 x3

J1 J2 J3 J12 J23

slide-20
SLIDE 20

linear connections singleton populations pairwise populations nonlinear connections

x1

x2 x3

J1 J2 J3 J12 J23

slide-21
SLIDE 21

Neural activity

slide-22
SLIDE 22

Neural activity

slide-23
SLIDE 23

Neural activity Neural encoding Information encoded

slide-24
SLIDE 24

Neural activity Neural encoding Information encoded

slide-25
SLIDE 25

Neural interactions Information interactions Neural encoding

slide-26
SLIDE 26

Neural interactions Probability distributions Information interactions Neural encoding

slide-27
SLIDE 27

Neural interactions Information interactions Neural encoding Example:

  • rientation
slide-28
SLIDE 28

min max True parameters

Nneurons no noise

Neural activity r Inferred parameters

Mean Variance

Time

Nparams = 1 Nneurons Nparams = 5

Network activity can implicitly perform inference

Raju and Pitkow 2016

slide-29
SLIDE 29

Simulated brain

Infer

time b

Encode

time r

Inferring inference

Decode

Message-passing parameters Interactions

Fit*

*within family

slide-30
SLIDE 30
  • ≠•!

True Learnt

Recovery results for simulated brain

Jij ij

G αβγ

αβγ

Message-passing parameters Interactions

slide-31
SLIDE 31

max

1 1

* * *

global min degenerate valley 1 degenerate valley 2 degenerate valley 2

Distance towards local minimum 1 Distance towards local minimum 2 Mean Squared Error

Analysis reveals degenerate family

  • f equivalent algorithms
slide-32
SLIDE 32

From simulated neural data we have recovered:

which variables interact Message-Passing algorithm Graphical model how they interact how the interactions are used how variables are encoded Representation

slide-33
SLIDE 33

Brain neural network Message passing nonlinearity

Applying message-passing to novel tasks

Apply to new graphical model structure Relax to novel neural network OR

slide-34
SLIDE 34

Next up: applying methods to real brains stimulus: orientation field recordings: V1 responses*

*not to same stimulus recordings from Tolias lab

slide-35
SLIDE 35

mementos:

  • Neurons can perform inference implicitly in a graphical

model distributed across a population.

  • New method to discover message-passing algorithms by

modeling transformations of decoded task variables

Brain World match

slide-36
SLIDE 36

xaqlab.com

Kaushik Lakshminarasimhan Qianli Yang Emin Orhan Aram Giahi-Saravani KiJung Yoon James Bridgewater Zhengwei Wu Saurabh Daptardar Rajkumar Vasudeva Raju

collaborators acknowledgements funding:

Alex Pouget Jeff Beck Dora Angelaki Andreas Tolias Jacob Reimer Fabian Sinz Alex Ecker Ankit Patel