Hidden Markov processes can explain complex sequencing rules of - - PowerPoint PPT Presentation

hidden markov processes can explain complex sequencing
SMART_READER_LITE
LIVE PREVIEW

Hidden Markov processes can explain complex sequencing rules of - - PowerPoint PPT Presentation

Hidden Markov processes can explain complex sequencing rules of birdsong: A statistical analysis and neural network modeling Kentaro Katahira 1,2,3 , Kenta Suzuki 3,4 , Kazuo Okanoya 1,2,3 , and Masato Okada 1,2,3 1. JST ERATO, Okanoya Emotional


slide-1
SLIDE 1

Hidden Markov processes can explain complex sequencing rules of birdsong: A statistical analysis and neural network modeling

Kentaro Katahira1,2,3, Kenta Suzuki3,4, Kazuo Okanoya1,2,3, and Masato Okada1,2,3

  • 1. JST ERATO, Okanoya Emotional Information Project,
  • 2. The University of Tokyo, 3. RIKEN Brain Science Institute,
  • 4. Saitama University
slide-2
SLIDE 2

Motivation

  • What are neural substrates for sequential behavior?

Generation Perception Learning

  • Speech
  • Playing music
  • Dancing

Sequential behavior

slide-3
SLIDE 3

Motivation

  • What are neural substrates for sequential behavior?

Birdsong

Generation Perception Learning

Syllable: a b c d Frequency

slide-4
SLIDE 4

Outline

  • 1. Introduction

– Neural substrates of birdsong – Neural network models

  • 2. Statistics of birdsong

– Higher-order history dependency

  • 3. Statistical models for birdsong
  • 4. Discussion

– Neural implementation – Future directions

slide-5
SLIDE 5

Neural activity pattern during singing

Hahnloser, Kozhevnikov and Fee, Nature, 2002 (Zebra finch)

slide-6
SLIDE 6

Feedforward chain hypothesis

  • Spikes propagate on feedforward chain network

Li & Greenside, Phys. Rev. E, 2006. Jin, Ramazanoglu, & Seung, J. Comput. Neurosci. 2007.

It is suitable for fixed sequences. But how about variable sequences?

Experimental evidences: Long & Fee, Nature, 2008; Long, Jin & Fee, Nature, 2010

slide-7
SLIDE 7

Song of Bengalese finch

  • Variable sequences including branching points
slide-8
SLIDE 8

Branching-chain hypothesis

inhibition (Jin, Phys Rev E, 2009)

a c b a b a c

Neuron Index Time

  • Mutual inhibition between branching chains
slide-9
SLIDE 9

Limitation of branching-chain model

  • The transition is a simple Markov process

– The present active chain depends only on the last active chain

Question: Syllable sequences of Bengalese finch songs are Markov processes?

Chain E Chain D Chain C Chain A Chain B Does not affect

?

slide-10
SLIDE 10

Outline

  • 1. Introduction

– Neural substrates of birdsong – Neural network models

  • 2. Statistics of birdsong

– Higher-order history dependency

  • 3. Statistical models for birdsong
  • 4. Discussion

– Neural implementation – Future directions

slide-11
SLIDE 11

Test of (first order) Markov assumption

Null hypothesis: The transition probability to next syllable does not depend on preceding syllable (Markov assumption) b d c e

0.495 0.408 0.097 Prob.

b d c e

0.385 0.422 0.193 Prob.

a χ2 goodness-of-fit test b a b

(For the case “a” precedes “b”)

Significant difference

→Second-order history dependency

slide-12
SLIDE 12

Result

b c a d a c

χ2(2) = 187.49, p < 0.0001

We found more than one significant second-order history dependency in all 16 birds.

(p < 0.01 with Bonferroni correction)

b c a d a c

0.13 0.55 0.31 0.99 < 0.01 < 0.01

slide-13
SLIDE 13

inhibition

A C B

Then,…

  • The branching-chain model is incorrect?

?

slide-14
SLIDE 14

Two possible mechanism for history dependency

Time (sec)

  • Freq. (Hz)

1.1 1.2 1.3 1.4 1.5 1.6 0.5 1 1.5 2 x 10

4a

c b d b a b c d

Chain 1 Chain 2 Chain 3 Chain 4

Hypothesis 1: Chain transition with higher-order dependency Hypothesis 2: Many-to-one mapping from chains to syllables

a b c d

Chain1 Chain2 Chain3 Chain4 Chain5

(Katahira, Okanoya and Okada, Biol. Cybern. 2007)

slide-15
SLIDE 15

However…

  • The neural activity data from HVC of singing

Bengalese finches are not available.

  • We examined two hypotheses based on song

data by using statistical models.

Bengalese finch Zebra finch

?

HVC HVC

Time (sec)
  • Freq. (Hz)
1.1 1.2 1.3 1.4 1.5 1.6 0.5 1 1.5 2 x 10 4
slide-16
SLIDE 16

Outline

  • 1. Introduction

– Neural substrates of birdsong – Neural network models

  • 2. Statistics of birdsong

– Higher-order history dependency

  • 3. Statistical models for birdsong
  • 4. Discussion

– Neural implementation – Future directions

slide-17
SLIDE 17

Feature extraction - Auditory features

Auditory features

  • Spectral entropy
  • Duration
  • Mean frequency

・ ・ ・ x1 x2 (c.f. Tchernichovski et al. 2000) Spectral entropy (z-score) Duration (z-score)

slide-18
SLIDE 18

Hidden Markov Model (HMM)

State 1 State 2 State 3

Hidden Observable

State 4

・ ・ ・ ・ ・ ・

a11 a12 a22 a23 a33 a24 a41

slide-19
SLIDE 19

State transition dynamics in HMM

1st order HMM: 0th order HMM (Gaussian mixture): 2nd order HMM:

slide-20
SLIDE 20

Relationship between two hypotheses and statistical models

→ 2nd order-HMM → 1st order-HMM

a b c d

Chain 1 Chain 2 Chain 3 Chain 4

Hypothesis 1: Chain transition with higher-order dependency

Hypothesis 2: Many-to-one mapping from chains to syllables

b a c d

Chain1 Chain2 Chain3 Chain4 Chain5

slide-21
SLIDE 21

Bayesian model selection

Model structure

  • L : Markov order (0,1,2)
  • K

: the number of hidden states Given data (auditory features): (→difficult to compute!) Model posterior: Marginal likelihood:

( : model parameter set)

Lower bound (variational free energy)

(can be computed by variational Bayes method)

Approximation

slide-22
SLIDE 22

Result – model selection (one bird)

“Best model structure”

  • With small number of states → 2nd order HMM
  • With large number of states → 1st order HMM

Number of states, K Lower bound on log-marginal likelihood

Better model

1st order HMM 2nd order HMM 0th order HMM

slide-23
SLIDE 23

Results – model selection, cross validation (averages over 16 birds)

Lower bound on log-marginal likelihood

Bound (z-score) 1st order HMM 2nd order HMM 0th order HMM

Predictive likelihood (cross validation)

1st order HMM 2nd order HMM 0th order HMM

slide-24
SLIDE 24

HMM learns many-to-one mapping

(Similar results were obtained for 30 syllables

  • f the 54 syllables where significant second-
  • rder dependency was found)

Many-to-one mapping from the states to a syllable “b”

slide-25
SLIDE 25

Outline

  • 1. Introduction

– Neural substrates of birdsong – Neural network models

  • 2. Statisticss of birdsong

– Higher-order history dependency

  • 3. Statistical models for birdsong
  • 4. Discussion

– Neural implementation – Future directions

slide-26
SLIDE 26

Summary of results

Time (sec)

  • Freq. (Hz)

1.1 1.2 1.3 1.4 1.5 1.6 0.5 1 1.5 2 x 10

4

a b c d

state1 state2 state3 state4 state5

a b c d

state1 state2 state3 state4 Many-to-one mapping – 1st HMM State transition with higher-order dependency - 2nd-order HMM

This mechanism is sufficient for Bengalese finch song

a c b d b

  • Bengalese finch songs have at least second-order

history dependency.

slide-27
SLIDE 27

Mapping onto neuroanatomy

  • HVC - hidden state (branch ⇔ state )
  • RA - auditory features of each syllable

(Katahira, Okanoya and Okada, 2007)

slide-28
SLIDE 28

Future directions (ongoing research)

  • How the brain can learn this representation?

– Analysis of development of song from a juvenile period. – Developing a network model with synaptic plasticity for learning the many-to-one mapping.

(e.g., Doya & Sejnowski, NIPS, 1995; Troyer & Doupe, J Neuropysiol, 2000; Fiete, Fee & Seung, J Neuropysiol,2007)

  • Applying HMMs to spike data recorded from songbird

(Katahira, Nishikawa, Okanoya & Okada, Neural Comput, 2010)

slide-29
SLIDE 29

Overbiew of our approach

Anatomy, Physiology Behavior

Time (sec)
  • Freq. (Hz)
1.1 1.2 1.3 1.4 1.5 1.6 0.5 1 1.5 2 x 10 4

Neural network model

Constraints

Statistical model

Parameter fitting, Model selection Mapping Constraints Support, Refinement