Incremental Parsing in Bounded Memory William Schuler Department of - - PowerPoint PPT Presentation

incremental parsing in bounded memory
SMART_READER_LITE
LIVE PREVIEW

Incremental Parsing in Bounded Memory William Schuler Department of - - PowerPoint PPT Presentation

Incremental Parsing in Bounded Memory William Schuler Department of Linguistics The Ohio State University September 16, 2010 William Schuler Incremental Parsing in Bounded Memory Motivation Goal: simple processing model, matches observations


slide-1
SLIDE 1

Incremental Parsing in Bounded Memory

William Schuler Department of Linguistics The Ohio State University September 16, 2010

William Schuler Incremental Parsing in Bounded Memory

slide-2
SLIDE 2

Motivation

Goal: simple processing model, matches observations about human memory

  • 1. bounded number of unconnected chunks

[Miller, 1956, Cowan, 2001] (subjects group stimuli into only 4 or so clusters)

William Schuler Incremental Parsing in Bounded Memory

slide-3
SLIDE 3

Motivation

Goal: simple processing model, matches observations about human memory

  • 1. bounded number of unconnected chunks

[Miller, 1956, Cowan, 2001] (subjects group stimuli into only 4 or so clusters)

  • 2. process rich syntax

[Chomsky and Miller, 1963] (center embedding: ‘if [neither [the man [the cop] saw] nor ...] then ...’) (cf. center recursion: ‘?? the malt [the rat [the cat chased] ate] ...’)

William Schuler Incremental Parsing in Bounded Memory

slide-4
SLIDE 4

Motivation

Goal: simple processing model, matches observations about human memory

  • 1. bounded number of unconnected chunks

[Miller, 1956, Cowan, 2001] (subjects group stimuli into only 4 or so clusters)

  • 2. process rich syntax

[Chomsky and Miller, 1963] (center embedding: ‘if [neither [the man [the cop] saw] nor ...] then ...’) (cf. center recursion: ‘?? the malt [the rat [the cat chased] ate] ...’)

  • 3. processing is incremental

[Sachs, 1967, Jarvella, 1971] (subjects can’t remember specific words earlier in sentence)

William Schuler Incremental Parsing in Bounded Memory

slide-5
SLIDE 5

Motivation

Goal: simple processing model, matches observations about human memory

  • 1. bounded number of unconnected chunks

[Miller, 1956, Cowan, 2001] (subjects group stimuli into only 4 or so clusters)

  • 2. process rich syntax

[Chomsky and Miller, 1963] (center embedding: ‘if [neither [the man [the cop] saw] nor ...] then ...’) (cf. center recursion: ‘?? the malt [the rat [the cat chased] ate] ...’)

  • 3. processing is incremental

[Sachs, 1967, Jarvella, 1971] (subjects can’t remember specific words earlier in sentence)

  • 4. processing is parallel, probabilistic

[Jurafsky, 1996, Hale, 2001, Levy, 2008] (probabilistic / info. theoretic measures correlate w. reading times)

William Schuler Incremental Parsing in Bounded Memory

slide-6
SLIDE 6

Motivation

Goal: simple processing model, matches observations about human memory In particular, we use a factored sequence model (dynamic Bayes net):

  • 1. random variables in Bayesian model are easily interpretable

(explicit estimation of speaker intent; cf. neural net)

William Schuler Incremental Parsing in Bounded Memory

slide-7
SLIDE 7

Motivation

Goal: simple processing model, matches observations about human memory In particular, we use a factored sequence model (dynamic Bayes net):

  • 1. random variables in Bayesian model are easily interpretable

(explicit estimation of speaker intent; cf. neural net)

  • 2. clear role of bounded working memory store

(random variable for each store element)

William Schuler Incremental Parsing in Bounded Memory

slide-8
SLIDE 8

Motivation

Goal: simple processing model, matches observations about human memory In particular, we use a factored sequence model (dynamic Bayes net):

  • 1. random variables in Bayesian model are easily interpretable

(explicit estimation of speaker intent; cf. neural net)

  • 2. clear role of bounded working memory store

(random variable for each store element)

  • 3. clear role of syntax

(grammar transform turns trees into chunks for store elements)

William Schuler Incremental Parsing in Bounded Memory

slide-9
SLIDE 9

Motivation

Goal: simple processing model, matches observations about human memory In particular, we use a factored sequence model (dynamic Bayes net):

  • 1. random variables in Bayesian model are easily interpretable

(explicit estimation of speaker intent; cf. neural net)

  • 2. clear role of bounded working memory store

(random variable for each store element)

  • 3. clear role of syntax

(grammar transform turns trees into chunks for store elements)

  • 4. fast enough to interact with real-time speech recognizer

(using cool engineering tricks: best-first / ‘lazy’ k-best search)

William Schuler Incremental Parsing in Bounded Memory

slide-10
SLIDE 10

Motivation

Goal: simple processing model, matches observations about human memory In particular, we use a factored sequence model (dynamic Bayes net):

  • 1. random variables in Bayesian model are easily interpretable

(explicit estimation of speaker intent; cf. neural net)

  • 2. clear role of bounded working memory store

(random variable for each store element)

  • 3. clear role of syntax

(grammar transform turns trees into chunks for store elements)

  • 4. fast enough to interact with real-time speech recognizer

(using cool engineering tricks: best-first / ‘lazy’ k-best search) Result is a nice platform for linguistic experimentation!

William Schuler Incremental Parsing in Bounded Memory

slide-11
SLIDE 11

Overview

Tutorial talk:

◮ Part I: Incremental Parsing

◮ bounded-memory sequence model ◮ connection to phrase structure ◮ coverage ◮ implementation/evaluation as performance model

◮ Part II: Extensions (Semantic Dependencies)

◮ preserving probabilistic dependencies in sequence model ◮ preserving semantic dependencies in sequence model ◮ interactive speech interpretation ◮ an analysis of non-local dependencies William Schuler Incremental Parsing in Bounded Memory

slide-12
SLIDE 12

Probabilistic Sequence Model

  • Hierarch. Hidden Markov Model [Murphy,Paskin’01]: bounded stack machine

s1

t− 1

s2

t− 1

s3

t− 1

DBN: circles=random variables (mem store elements), arcs=dependencies Elements hold hypoth. stacked-up incomplete constituents, dep. on parent (incomplete constituent: e.g. S/VP = sentence lacking verb phrase to come)

William Schuler Incremental Parsing in Bounded Memory

slide-13
SLIDE 13

Probabilistic Sequence Model

  • Hierarch. Hidden Markov Model [Murphy,Paskin’01]: bounded stack machine

s1

t− 1

s2

t− 1

s3

t− 1

xt−

1

DBN: circles=random variables (mem store elements), arcs=dependencies Elements hold hypoth. stacked-up incomplete constituents, dep. on parent (incomplete constituent: e.g. S/VP = sentence lacking verb phrase to come) Hypothesized mem elements generate observations: words / acoust. features

William Schuler Incremental Parsing in Bounded Memory

slide-14
SLIDE 14

Probabilistic Sequence Model

  • Hierarch. Hidden Markov Model [Murphy,Paskin’01]: bounded stack machine

s1

t− 1

s2

t− 1

s3

t− 1

xt−

1

r3

t

r2

t

r1

t

Elements in memory store may be composed (reduced) w. element above Probability depends on dependent vars (e.g. Det, Noun reduce to NP)

William Schuler Incremental Parsing in Bounded Memory

slide-15
SLIDE 15

Probabilistic Sequence Model

  • Hierarch. Hidden Markov Model [Murphy,Paskin’01]: bounded stack machine

s1

t− 1

s2

t− 1

s3

t− 1

xt−

1

r3

t

r2

t

r1

t

s1

t

s2

t

s3

t

(Non-)reduced elements carry forward or transition (e.g. NP becomes S/VP)

William Schuler Incremental Parsing in Bounded Memory

slide-16
SLIDE 16

Probabilistic Sequence Model

  • Hierarch. Hidden Markov Model [Murphy,Paskin’01]: bounded stack machine

s1

t− 1

s2

t− 1

s3

t− 1

xt−

1

r3

t

r2

t

r1

t

s1

t

s2

t

s3

t

xt

(Non-)reduced elements carry forward or transition (e.g. NP becomes S/VP) Transitioned elements may be expanded again (e.g. S/VP expands to Verb)

William Schuler Incremental Parsing in Bounded Memory

slide-17
SLIDE 17

Probabilistic Sequence Model

  • Hierarch. Hidden Markov Model [Murphy,Paskin’01]: bounded stack machine

. . . . . . . . . . . . r3

t− 1

r2

t− 1

r1

t− 1

s1

t− 1

s2

t− 1

s3

t− 1

xt−

1

r3

t

r2

t

r1

t

s1

t

s2

t

s3

t

xt r3

t+ 1

r2

t+ 1

r1

t+ 1

s1

t+ 1

s2

t+ 1

s3

t+ 1

xt+

1

(Non-)reduced elements carry forward or transition (e.g. NP becomes S/VP) Transitioned elements may be expanded again (e.g. S/VP expands to Verb) Process continues through time

William Schuler Incremental Parsing in Bounded Memory

slide-18
SLIDE 18

Probabilistic Sequence Model

  • Hierarch. Hidden Markov Model [Murphy,Paskin’01]: bounded stack machine

. . . . . . . . . . . . r3

t− 1

r2

t− 1

r1

t− 1

s1

t− 1

s2

t− 1

s3

t− 1

xt−

1

r3

t

r2

t

r1

t

s1

t

s2

t

s3

t

xt r3

t+ 1

r2

t+ 1

r1

t+ 1

s1

t+ 1

s2

t+ 1

s3

t+ 1

xt+

1

Alternate hypotheses (memory store configurations) compete w. each other: ˆ s1..D

1..T def

= argmax

s1..D

1..T

T

  • t=1

PθY(s1..D

t

| s1..D

t− 1 ) · PθX(xt | s1..D t

)

William Schuler Incremental Parsing in Bounded Memory

slide-19
SLIDE 19

Probabilistic Sequence Model

  • Hierarch. Hidden Markov Model [Murphy,Paskin’01]: bounded stack machine

. . . . . . . . . . . . r3

t− 1

r2

t− 1

r1

t− 1

s1

t− 1

s2

t− 1

s3

t− 1

xt−

1

r3

t

r2

t

r1

t

s1

t

s2

t

s3

t

xt r3

t+ 1

r2

t+ 1

r1

t+ 1

s1

t+ 1

s2

t+ 1

s3

t+ 1

xt+

1

PθY(s1..D

t

| s1..D

t-1 ) =

  • r1..D

t

PθReduce(r1..D

t

| s1..D

t-1 ) · PθShift(s1..D t

| r1..D

t

s1..D

t-1 ) def

=

  • r1..D

t

D

  • d=1

P

θR,d(rd t | rd+ 1 t

sd

t-1sd-1 t-1 ) · P θS,d(sd t | rd+ 1 t

rd

t sd t-1sd-1 t

)

William Schuler Incremental Parsing in Bounded Memory

slide-20
SLIDE 20

Probabilistic Sequence Model

  • Hierarch. Hidden Markov Model [Murphy,Paskin’01]: bounded stack machine

. . . . . . . . . . . . r3

t− 1

r2

t− 1

r1

t− 1

s1

t− 1

s2

t− 1

s3

t− 1

xt−

1

r3

t

r2

t

r1

t

s1

t

s2

t

s3

t

xt r3

t+ 1

r2

t+ 1

r1

t+ 1

s1

t+ 1

s2

t+ 1

s3

t+ 1

xt+

1

Natural independence assumptions:

◮ time-order: each store depends on immediately previous store ◮ depth-order: incomplete constituents separated by yet-unknown struct

William Schuler Incremental Parsing in Bounded Memory

slide-21
SLIDE 21

Probabilistic Sequence Model

Add interactive semantics — simply factor HHMM states:

. . . . . . . . . . . . r3

t− 1 es3

t − 1 fr3 t − 1

r2

t− 1 es2

t − 1 fr2 t − 1

r1

t− 1 es1

t − 1 fr1 t − 1

s1

t− 1 es1

t − 1 qs1 t − 1

s2

t− 1 es2

t − 1 qs2 t − 1

s3

t− 1 es3

t − 1 qs3 t − 1

xt−

1

r3

t

es3

t

fr3

t

r2

t

es2

t

fr2

t

r1

t

es1

t

fr1

t

s1

t

es1

t

qs1

t

s2

t

es2

t

qs2

t

s3

t

es3

t

qs3

t

xt

— factor r , s into interdependent syntactic (q/f ) and referential (e) states: rd

t def

= esd

t , frd t

sd

t def

= esd

t , qsd t William Schuler Incremental Parsing in Bounded Memory

slide-22
SLIDE 22

Probabilistic Sequence Model

Add interactive semantics — simply factor HHMM states:

. . . . . . . . . . . . r3

t− 1 es3

t − 1 fr3 t − 1

r2

t− 1 es2

t − 1 fr2 t − 1

r1

t− 1 es1

t − 1 fr1 t − 1

s1

t− 1 es1

t − 1 qs1 t − 1

s2

t− 1 es2

t − 1 qs2 t − 1

s3

t− 1 es3

t − 1 qs3 t − 1

xt−

1

r3

t

es3

t

fr3

t

r2

t

es2

t

fr2

t

r1

t

es1

t

fr1

t

s1

t

es1

t

qs1

t

s2

t

es2

t

qs2

t

s3

t

es3

t

qs3

t

xt

— factor r , s into interdependent syntactic (q/f ) and referential (e) states:

◮ incomplete syntactic states: e.g. q = S/VP (with fr as a reduce flag)

William Schuler Incremental Parsing in Bounded Memory

slide-23
SLIDE 23

Probabilistic Sequence Model

Add interactive semantics — simply factor HHMM states:

. . . . . . . . . . . . r3

t− 1 es3

t − 1 fr3 t − 1

r2

t− 1 es2

t − 1 fr2 t − 1

r1

t− 1 es1

t − 1 fr1 t − 1

s1

t− 1 es1

t − 1 qs1 t − 1

s2

t− 1 es2

t − 1 qs2 t − 1

s3

t− 1 es3

t − 1 qs3 t − 1

xt−

1

r3

t

es3

t

fr3

t

r2

t

es2

t

fr2

t

r1

t

es1

t

fr1

t

s1

t

es1

t

qs1

t

s2

t

es2

t

qs2

t

s3

t

es3

t

qs3

t

xt

— factor r , s into interdependent syntactic (q/f ) and referential (e) states:

◮ incomplete syntactic states: e.g. q = S/VP (with fr as a reduce flag) ◮ incomplete referential states: e.g. e ={icoling, inaacl} (concept set/reln)

William Schuler Incremental Parsing in Bounded Memory

slide-24
SLIDE 24

Connecting Phrase Structure to Sequence Model

Sequences of memory stores correspond directly to familiar phrase structure: (trees from Penn Treebank, binarized around head rules)

S NP NP JJ strong NN demand PP IN for NP NPpos NNP NNP new NNP NNP york NNP city POS ’s NNS JJ general NNS NN

  • bligation

NNS bonds VP VBN VBN propped PRT up NP DT the NN JJ municipal NN market

Phrase structure allows nested expressions: ‘prop ... up,’ ‘if ... then ...’ Bounded memory requires flatter, more memory-efficient representation...

William Schuler Incremental Parsing in Bounded Memory

slide-25
SLIDE 25

Connecting Phrase Structure to Sequence Model

‘Right-corner transform’ map right-embedded sequence → left-embedded seq. (allows new constituents to be immediately composed) Limited: center embedding still consumes memory [Chomsky and Miller, 1963]

S S/NN S/NN S/NP S/VP NP NP/NNS NP/NNS NP/NNS NP/NP NP/PP NP NP/NN JJ strong NN demand IN for NPpos NPpos/POS NNP NNP/NNP NNP/NNP NNP new NNP york NNP city POS ’s JJ general NN

  • bligation

NNS bonds VBN VBN/PRT VBN propped PRT up DT the JJ municipal NN market

William Schuler Incremental Parsing in Bounded Memory

slide-26
SLIDE 26

Connecting Phrase Structure to Sequence Model

Transform is simple — three cases on any right-embedded sequence:

(η, ι are paths of 0:left/1:right)

Beginning case:

cη cη·0 α cη·1 β

cη cη/cη·1 cη·0 α β (∼ CCG type raising)

Middle case:

cη α cη·ι cη·ι·0 β cη·ι·1 γ

cη cη/cη·ι·1 cη/cη·ι α cη·ι·0 β γ (∼ fwd. composition)

Ending case:

cη α cη·ι aη·ι

cη cη/cη·ι α cη·ι aη·ι (∼ fwd. application)

William Schuler Incremental Parsing in Bounded Memory

slide-27
SLIDE 27

Connecting Phrase Structure to Sequence Model

Align trees to HHMM grid, train variables on incomplete constit. ‘chunks’:

d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 t=7 t=8 t=9 t=10 t=11 t=12 t=13 t=14 t=15 strong demand for new york city ’s general

  • bligation

bonds propped up the municipal market − − − − − − − − − − − − − − − − − − − NNP/NNP NNP/NNP NPpos/POS − − − − VBN/PRT − − − − NP/NN NP/PP NP/NP NP/NP NP/NP NP/NP NP/NNS NP/NNS NP/NNS S/VP S/VP S/NP S/NN S/NN William Schuler Incremental Parsing in Bounded Memory

slide-28
SLIDE 28

Connecting Phrase Structure to Sequence Model

Align trees to HHMM grid, train variables on incomplete constit. ‘chunks’:

d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 t=7 t=8 t=9 t=10 t=11 t=12 t=13 t=14 t=15 strong demand for new york city ’s general

  • bligation

bonds propped up the municipal market − − − − − − − − − − − − − − − − − − − NNP/NNP NNP/NNP NPpos/POS − − − − VBN/PRT − − − − NP/NN NP/PP NP/NP NP/NP NP/NP NP/NP NP/NNS NP/NNS NP/NNS S/VP S/VP S/NP S/NN S/NN

Unlike incomplete constit. chunks from left-corner transform [Johnson ’98]:

◮ right-corner gives fixed struct w/in chunk, unknown struct between ◮ time-order traversal gives bottom-up (compositional) semantics

William Schuler Incremental Parsing in Bounded Memory

slide-29
SLIDE 29

Connecting Phrase Structure to Sequence Model

Align trees to HHMM grid, train variables on incomplete constit. ‘chunks’:

d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 t=7 t=8 t=9 t=10 t=11 t=12 t=13 t=14 t=15 strong demand for new york city ’s general

  • bligation

bonds propped up the municipal market − − − − − − − − − − − − − − − − − − − NNP/NNP NNP/NNP NPpos/POS − − − − VBN/PRT − − − − NP/NN NP/PP NP/NP NP/NP NP/NP NP/NP NP/NNS NP/NNS NP/NNS S/VP S/VP S/NP S/NN S/NN

Unlike full Combinatory Categorial Grammar [Steedman, 2000]:

◮ not pure lexical: semantics can come from grammar rules/‘constructions’ ◮ no commitment to set of combinators

William Schuler Incremental Parsing in Bounded Memory

slide-30
SLIDE 30

Coverage

Penn Treebank WSJ coverage experiment indicates 3- to 4-element store [Schuler et al., 2008, Schuler et al., 2010] stack memory capacity sentences coverage no stack memory 127 0.32% 1 stack element 3,496 8.78% 2 stack elements 25,909 65.05% 3 stack elements 38,902 97.67% 4 stack elements 39,816 99.96% 5 stack elements 39,832 100.00% TOTAL 39,832 100.00% (percent coverage of transformed Treebank sections 2–21 w /

  • punctuation)

Good, because 3 to 4 elements supposed to be our limit [Cowan, 2001]

William Schuler Incremental Parsing in Bounded Memory

slide-31
SLIDE 31

Coverage

Penn Treebank WSJ coverage experiment indicates 3- to 4-element store [Schuler et al., 2008, Schuler et al., 2010] stack memory capacity sentences coverage no stack memory 127 0.32% 1 stack element 3,496 8.78% 2 stack elements 25,909 65.05% 3 stack elements 38,902 97.67% 4 stack elements 39,816 99.96% 5 stack elements 39,832 100.00% TOTAL 39,832 100.00% (percent coverage of transformed Treebank sections 2–21 w /

  • punctuation)

Good, because 3 to 4 elements supposed to be our limit [Cowan, 2001] Stronger for Switchboard (spontaneous speech): 92.1% at 2, 99.5% at 3.

William Schuler Incremental Parsing in Bounded Memory

slide-32
SLIDE 32

Coverage

Penn Treebank WSJ coverage experiment indicates 3- to 4-element store Significant (p<.0001) divergence from corpus randomly generated from pcfg

5000 10000 15000 20000 25000 2 4 6 8 10 "./genmodel/rand.pcfg.wsjnp.depths" using 2:1 "./genmodel/all.wsjnp.depths" using 2:1

William Schuler Incremental Parsing in Bounded Memory

slide-33
SLIDE 33

Coverage

Here’s one of the 16 depth-five sentences in the (40,000 sent.) corpus:

S

  • n Asia-Pacific prosperity

S SBAR if S S America can VP VP keep up NP NP the present situation NP her markets open for another 15 years ... and Japan can grow and ... then ...

Does indeed seem like a lot to remember

William Schuler Incremental Parsing in Bounded Memory

slide-34
SLIDE 34

Evaluation as a Performance Model

Setting D=4 should allow nearly complete coverage. But strict memory bounds may introduce spurious local ambiguity...

William Schuler Incremental Parsing in Bounded Memory

slide-35
SLIDE 35

Evaluation as a Performance Model

Setting D=4 should allow nearly complete coverage. But strict memory bounds may introduce spurious local ambiguity... Bounded memory forces model to anticipate nesting demands: (a) allow modification to ‘city’ NP / (b) leave room for future embedding:

a) d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 t=7 s t r

  • n

g d e m a n d f

  • r

n e w y

  • r

k c i t y − − − − − − − − − − N N P / N N P N N P / N N P N P ( c i t y ) − N P / N N N P / P P N P / N P N P / N P N P / N P N P ( d e m . ) / N P ( ? ) b) d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 t=7 s t r

  • n

g d e m a n d f

  • r

n e w y

  • r

k c i t y − − − − − − − − − − − − − N P / N N N P / P P N P / N P N P / N N P N P / N N P N P ( d e m . )

William Schuler Incremental Parsing in Bounded Memory

slide-36
SLIDE 36

Evaluation as a Performance Model

Setting D=4 should allow nearly complete coverage. But strict memory bounds may introduce spurious local ambiguity... Bounded memory forces model to anticipate nesting demands: (a) allow modification to ‘city’ NP / (b) leave room for future embedding:

a) d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 t=7 s t r

  • n

g d e m a n d f

  • r

n e w y

  • r

k c i t y − − − − − − − − − − N N P / N N P N N P / N N P N P ( c i t y ) − N P / N N N P / P P N P / N P N P / N P N P / N P N P ( d e m . ) / N P ( ? ) b) d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 t=7 s t r

  • n

g d e m a n d f

  • r

n e w y

  • r

k c i t y − − − − − − − − − − − − − N P / N N N P / P P N P / N P N P / N N P N P / N N P N P ( d e m . )

As defined, model makes this prediction probabilistically (parsing strategy is optionally ‘arc-eager’ [Abney and Johnson, 1991]). But, could be overwhelmed with memory-management ambiguity...

William Schuler Incremental Parsing in Bounded Memory

slide-37
SLIDE 37

Evaluation as a Performance Model

Implemented using ‘Modelblocks’ C++ templates for sequence models: PθY(s1..D

t

| s1..D

t-1 ) =

  • r1..D

t

PθReduce(r1..D

t

| s1..D

t-1 ) · PθShift(s1..D t

| r1..D

t

s1..D

t-1 ) def

=

  • r1..D

t

D

  • d=1

P

θR,d(rd t | rd+ 1 t

sd

t-1sd-1 t-1 ) · P θS,d(sd t | rd+ 1 t

rd

t sd t-1sd-1 t

)

class YModel { public: RModel mR; SModel mS; LogProb setIterProb ( Y::AIterator<LogProb>& y, const S& sP, const X& x, bool b1, int& a ) const { LogProb pr; pr = mR.setIterProb ( y.first, sP, b1, a ); pr *= mS.setIterProb ( y.second, y.first, sP, a ); return pr; } // ... }; William Schuler Incremental Parsing in Bounded Memory

slide-38
SLIDE 38

Evaluation as a Performance Model

Implemented using ‘Modelblocks’ C++ templates for sequence models: PθY(s1..D

t

| s1..D

t-1 ) =

  • r1..D

t

PθReduce(r1..D

t

| s1..D

t-1 ) · PθShift(s1..D t

| r1..D

t

s1..D

t-1 ) def

=

  • r1..D

t

D

  • d=1

P

θR,d(rd t | rd+ 1 t

sd

t-1sd-1 t-1 ) · P θS,d(sd t | rd+ 1 t

rd

t sd t-1sd-1 t

)

class RModel { public: RdModel mRd; LogProb setIterProb ( R::AIterator<LogProb>& r, const S& sP, bool b1, int& a ) const { LogProb pr; pr = mRd.setIterProb ( r[3], 4, Rd(sP.second), sP.first[3], sP.first[2], b1, a ); pr *= mRd.setIterProb ( r[2], 3, Rd(r[3]), sP.first[2], sP.first[1], b1, a ); pr *= mRd.setIterProb ( r[1], 2, Rd(r[2]), sP.first[1], sP.first[0], b1, a ); pr *= mRd.setIterProb ( r[0], 1, Rd(r[1]), sP.first[0], Sd TOP, b1, a ); return pr; } }; William Schuler Incremental Parsing in Bounded Memory

slide-39
SLIDE 39

Evaluation as a Performance Model

Implemented using ‘Modelblocks’ C++ templates for sequence models: PθY(s1..D

t

| s1..D

t-1 ) =

  • r1..D

t

PθReduce(r1..D

t

| s1..D

t-1 ) · PθShift(s1..D t

| r1..D

t

s1..D

t-1 ) def

=

  • r1..D

t

D

  • d=1

P

θR,d(rd t | rd+ 1 t

sd

t-1sd-1 t-1 ) · P θS,d(sd t | rd+ 1 t

rd

t sd t-1sd-1 t

)

class SModel { public: SdModel mSd; LogProb setIterProb(S::AIterator<LogProb>& s, const R::AIterator<LogProb>& r, const S& sP, int& a) const { LogProb pr; pr = mSd.setIterProb ( s.first[0], 1, Rd(r[1]), Rd(r[0]), sP.first[0], Sd TOP, a ); pr *= mSd.setIterProb ( s.first[1], 2, Rd(r[2]), Rd(r[1]), sP.first[1], Sd(s.first[0]), a ); pr *= mSd.setIterProb ( s.first[2], 3, Rd(r[3]), Rd(r[2]), sP.first[2], Sd(s.first[1]), a ); pr *= mSd.setIterProb ( s.first[3], 4, sP.second, Rd(r[3]), sP.first[3], Sd(s.first[2]), a ); pr *= ( G(s.first[3].second)!=G BOT && G(s.first[3].second).getTerm()!=B 1 ) ? mSd.mGe.setIterProb ( s.second, 5, G(s.first[3].second), a ) : mG BOT.setIterProb ( s.second, a ); return pr; } }; William Schuler Incremental Parsing in Bounded Memory

slide-40
SLIDE 40

Evaluation as a Performance Model

Implemented using ‘Modelblocks’ C++ templates for sequence models: PθY(s1..D

t

| s1..D

t-1 ) =

  • r1..D

t

PθReduce(r1..D

t

| s1..D

t-1 ) · PθShift(s1..D t

| r1..D

t

s1..D

t-1 ) def

=

  • r1..D

t

D

  • d=1

P

θR,d(rd t | rd+ 1 t

sd

t-1sd-1 t-1 ) · P θS,d(sd t | rd+ 1 t

rd

t sd t-1sd-1 t

) Factored model is compiled into HMM Viterbi recognizer using template:

#include "TextObsVars.h" #include "HHMMLangModel-gf.h" #include "TextObsModel.h" #include "HHMMParser.h" int main (int nArgs, char* argv[]) { HMM Viterbi MLS<YModel,XModel,S,R> ( nArgs, argv ); }

Optimized with ‘lazy’ k-best search across variables at each time step.

William Schuler Incremental Parsing in Bounded Memory

slide-41
SLIDE 41

Evaluation as a Performance Model

Parser still accurate at narrow beam widths [Miller and Schuler, 2010]:

70 72 74 76 78 80 82 84 100 200 300 400 500 Labeled F-Score Beam Width

Accuracy on WSJ sect 23 using beam width 15, 20, 25, 50, 100, 250, 500.

William Schuler Incremental Parsing in Bounded Memory

slide-42
SLIDE 42

Evaluation as a Performance Model

Speed competitive with CKY parsing [Miller and Schuler, 2010]:

2 4 6 8 10 12 14 10 20 30 40 50 60 70 Seconds per sentence Sentence Length CKY HHMM

HHMM beam 20 (74%F) vs. ‘vanilla’ CKY [Klein & Manning,’03] (71%F).

William Schuler Incremental Parsing in Bounded Memory

slide-43
SLIDE 43

Evaluation as a Performance Model

Does optionally arc-eager strategy support surprisal on reading time? [Hale, 2001, Roark et al., 2009]

William Schuler Incremental Parsing in Bounded Memory

slide-44
SLIDE 44

Evaluation as a Performance Model

Does optionally arc-eager strategy support surprisal on reading time? [Hale, 2001, Roark et al., 2009] Evaluated correlation with reading times on [Bachrach et al., 2009]: “joe was a big bear of a man six feet six inches tall and barrel-chested” “when he fell off the moon tower the tremor his fall caused seemed not unlike an earthquake to people who lived by”

William Schuler Incremental Parsing in Bounded Memory

slide-45
SLIDE 45

Evaluation as a Performance Model

Does optionally arc-eager strategy support surprisal on reading time? [Hale, 2001, Roark et al., 2009] Evaluated correlation with reading times on [Bachrach et al., 2009]: “joe was a big bear of a man six feet six inches tall and barrel-chested” “when he fell off the moon tower the tremor his fall caused seemed not unlike an earthquake to people who lived by” Factors: Factor Description Expected word order read faster as story goes on

  • neg. slope

reciprocal length read longer words slower

  • pos. slope

unigram freq. read common words faster

  • neg. slope

bigram prob. read common bigrams faster

  • neg. slope

embedding diff. read embedded phrases slower

  • pos. slope

entropy reduction read perplexing words slower

  • pos. slope

surprisal read surprising words slower

  • pos. slope

William Schuler Incremental Parsing in Bounded Memory

slide-46
SLIDE 46

Evaluation as a Performance Model

Yes! Results of linear mixed-effect modeling [Wu et al., 2010]: Factor Coefficient

  • Std. Err.

t-value (Intercept)

  • 9.340·10−3

5.347·10−2

  • 0.175

word order

  • 3.746·10−5

7.808·10−6

  • 4.797∗

reciprocal length

  • 2.002·10−2

1.635·10−2

  • 1.225

unigram freq.

  • 8.090·10−2

3.690·10−1

  • 0.219

bigram prob.

  • 2.074·10+0

8.132·10−1

  • 2.551∗

embedding diff. 9.390·10−3 3.268·10−3 2.873∗ entropy reduction 2.753·10−2 6.792·10−3 4.052∗ surprisal 3.950·10−3 3.452·10−4 11.442∗ Significance (indicated by ∗) is reported at p < 0.05.

William Schuler Incremental Parsing in Bounded Memory

slide-47
SLIDE 47

Evaluation as a Performance Model

Incomplete constituents also provide nice account of disfluency [Miller and Schuler, 2008, Miller, 2009]

d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 t=7 t=8 t=9 t h e y s a i d s h e h a d a s h e h a d s

  • m

e t r

  • u

b l e − − − − − − − − − − − − S / V P S / N P − − − − − S / V P S / S S / S S / S S / S S / V P S / N P S / N N

‘She had a’ is reparandum; intended to be replaced w. ‘she had some trouble.’ Here, incomplete constituent can be fluent (≈ conjunction) until repair.

William Schuler Incremental Parsing in Bounded Memory

slide-48
SLIDE 48

Overview

Tutorial talk:

◮ Part I: Incremental Parsing

◮ bounded-memory sequence model ◮ connection to phrase structure ◮ coverage ◮ implementation/evaluation as performance model

◮ Part II: Extensions (Semantic Dependencies)

◮ preserving probabilistic dependencies in sequence model ◮ preserving semantic dependencies in sequence model ◮ interactive speech interpretation ◮ an analysis of non-local dependencies William Schuler Incremental Parsing in Bounded Memory

slide-49
SLIDE 49

Preserving Dependencies in Sequence Model

Transformation on trees is ok for syntax, but to study effects of lexicalization / selectional restrictions / semantics, need to preserve dependencies...

William Schuler Incremental Parsing in Bounded Memory

slide-50
SLIDE 50

Preserving Dependencies in Sequence Model

Transformation on trees is ok for syntax, but to study effects of lexicalization / selectional restrictions / semantics, need to preserve dependencies... Define factored HMM probabilities in terms of original CFG: 1) Start by decomposing yield into yields of left and right children ¯ xη = ¯ xη·0 · ¯ xη·1 2) Equivalently, decompose into yield of incomplete & complete constituent ¯ xη = (¯ xη/¯ xη·ι) · ¯ xη·ι for all ι ∈ 1+

William Schuler Incremental Parsing in Bounded Memory

slide-51
SLIDE 51

Preserving Dependencies in Sequence Model

3) Using these yields, define recurrences analogous to transform rules: PθIns(G)(¯ xη | cη) =

  • ι∈1+,¯

xηι,cηι

PθIC(G)(¯ xη/¯ xηι, cηι | cη) · PθIns(G)(¯ xηι | cηι) analogous to ‘end case’: (forward application) cη ¯ xη/¯ xηι cηι ¯ xηι

cη cη/ cηι ¯ xη/¯ xηι cηι ¯ xηι

William Schuler Incremental Parsing in Bounded Memory

slide-52
SLIDE 52

Preserving Dependencies in Sequence Model

3) Using these yields, define recurrences analogous to transform rules: PθIC(G)(¯ xη/¯ xηι1, cηι1 | cη) =

  • ¯

xηι,cηι

PθIC(G)(¯ xη/¯ xηι, cηι | cη) ·

  • ¯

xηι0,cηι0

PθG(cηι → cηι0 cηι1) · PθIns(G)(¯ xηι0 | cηι0) analogous to ‘middle case’: (forward composition) cη ¯ xη/¯ xηι cηι cηι0 ¯ xηι0 cηι1 ¯ xηι1

cη cη/ cηι1 cη/ cηι ¯ xη/¯ xηι cηι0 ¯ xηι0 ¯ xηι1

William Schuler Incremental Parsing in Bounded Memory

slide-53
SLIDE 53

Preserving Dependencies in Sequence Model

3) Using these yields, define recurrences analogous to transform rules: PθIC(G)(¯ xη/¯ xη1, cη1 | cη) =

  • ¯

xη0,cη0

PθG(cη → cη0 cη1) · PθIns(G)(¯ xη0 | cη0) analogous to ‘begin case’: (type raising) cη cη0 ¯ xη0 cη1 ¯ xη1

cη cη/ cη1 cη0 ¯ xη0 ¯ xη1

William Schuler Incremental Parsing in Bounded Memory

slide-54
SLIDE 54

Preserving Dependencies in Sequence Model

4) Connect each incomplete constituent as leftmost descendant of previous: PθFwd((¯ xηd /¯ xηdιd )D

d=1, (cηd )D d=1, (cηd ιd)D d=1) = D

  • d=1

PθG-rl∗

,d (cηd−1ιd−1

→ cηd ...) · PθIC(G)(¯ xηd /¯ xηdιd, cηdιd | cηd )

William Schuler Incremental Parsing in Bounded Memory

slide-55
SLIDE 55

Preserving Dependencies in Sequence Model

4) Connect each incomplete constituent as leftmost descendant of previous: PθFwd((¯ xηd /¯ xηdιd )D

d=1, (cηd )D d=1, (cηd ιd)D d=1) = D

  • d=1

PθG-rl∗

,d (cηd−1ιd−1

→ cηd ...) · PθIC(G)(¯ xηd /¯ xηdιd, cηdιd | cηd ) with leftmost-descendant terms defined by value iteration [Bellman, 1957]: EθG-rl∗

,d (cη

→cη0 ...) =

  • cη1

PθG-r,d(cη →cη0 cη1) EθG-rl∗

,d(cη

k

→cη0k 0 ...) =

  • cη0k ,cη0k 1

EθG-rl∗

,d(cη

k− 1

→ cη0k ...) · PθG-l,d(cη0k →cη0k 0 cη0k1) EθG-rl∗

,d(cη

→cηι ...) =

  • k=0

EθG-rl∗

,d(cη

k

→cηι ...) EθG-rl∗

,d(cη

+

→cηι ...) = EθG-rl∗

,d(cη

→cηι ...) − EθG-rl∗

,d(cη

→cηι ...)

William Schuler Incremental Parsing in Bounded Memory

slide-56
SLIDE 56

Preserving Dependencies in Sequence Model

5) Then, associating store variables with incomplete constituent categories... sd

t def

= cηd , cηd ιd s.t. ιd ∈ 1+, ηd ∈ ηd−1ιd−10+ rd

t def

= cηd , frd

t

s.t. frd

t = ¯

xηd = cηd (reduction or not)

William Schuler Incremental Parsing in Bounded Memory

slide-57
SLIDE 57

Preserving Dependencies in Sequence Model

5) Then, associating store variables with incomplete constituent categories... sd

t def

= cηd , cηd ιd s.t. ιd ∈ 1+, ηd ∈ ηd−1ιd−10+ rd

t def

= cηd , frd

t

s.t. frd

t = ¯

xηd = cηd (reduction or not) divide incomplete constituents into components of HMM recurrence... PθFwd(x1..t s1..D

t

) =

  • η1..D,ι1..D

PθFwd((¯ xηd /¯ xηdιd )D

d=1, (cηd )D d=1, (cηd ιd)D d=1)

=

  • s1..D

t−1

PθFwd(x1..t−

1 s1..D t− 1 ) · PθY(s1..D t

| s1..D

t− 1 ) · PθX(xt | s1..D t

)

William Schuler Incremental Parsing in Bounded Memory

slide-58
SLIDE 58

Preserving Dependencies in Sequence Model

5) Then, associating store variables with incomplete constituent categories... sd

t def

= cηd , cηd ιd s.t. ιd ∈ 1+, ηd ∈ ηd−1ιd−10+ rd

t def

= cηd , frd

t

s.t. frd

t = ¯

xηd = cηd (reduction or not) divide incomplete constituents into components of HMM recurrence... PθFwd(x1..t s1..D

t

) =

  • η1..D,ι1..D

PθFwd((¯ xηd /¯ xηdιd )D

d=1, (cηd )D d=1, (cηd ιd)D d=1)

=

  • s1..D

t−1

PθFwd(x1..t−

1 s1..D t− 1 ) · PθY(s1..D t

| s1..D

t− 1 ) · PθX(xt | s1..D t

) using the hidden state model already defined: PθY(s1..D

t

| s1..D

t-1 ) =

  • r1..D

t

PθReduce(r1..D

t

| s1..D

t-1 ) · PθShift(s1..D t

| r1..D

t

s1..D

t-1 ) def

=

  • r1..D

t

D

  • d=1

P

θR,d(rd t | rd+ 1 t

sd

t-1sd-1 t-1 ) · P θS,d(sd t | rd+ 1 t

rd

t sd t-1sd-1 t

)

William Schuler Incremental Parsing in Bounded Memory

slide-59
SLIDE 59

Preserving Dependencies in Sequence Model

Model uses reduction states rd

t to ensure only one transition per time step

(all reductions below some depth, none above; transition is at crossover) P

θR,d(rd t | rd+ 1 t

sd

t− 1sd− 1 t− 1 ) def

= if frd+

1 t

=0 : rd

t =r⊥

if frd+

1 t

=1 : PθR-r,d(rd

t | rd+ 1 t

sd

t− 1 sd− 1 t− 1 )

P

θS,d(sd t | rd+ 1 t

rd

t sd t− 1sd− 1 t

) def =    if frd+

1 t

=1, frd

t =1 : PθS-e,d (sd

t | sd− 1 t

) if frd+

1 t

=1, frd

t =0 : PθS-t,d(sd

t | rd+ 1 t

rd

t sd t− 1sd− 1 t

) if frd+

1 t

=0, frd

t =0 : sd

t =sd t− 1

Models now defined on CFG probs + chopped-up left-descendant terms... Expansion probabilities simply contribute new left-descendant term: PθS-e,d(cηι, c′

ηι | −, cη) def

= EθG-rl∗

,d(cη

→ cηι ...) · xηι =c′

ηι = cηι

William Schuler Incremental Parsing in Bounded Memory

slide-60
SLIDE 60

Preserving Dependencies in Sequence Model

Models now defined on CFG probs + chopped-up left-descendant terms... Reductions divide left-descendant prob into 0/ >0 left-expansions: PθR-r,d (rd

t | rd+ 1 t

sd

t− 1sd− 1 t− 1 ) def

= if crd+

1 t

=xt : rd

t =r⊥

if crd+

1 t

=xt : PθR-r,d(rd

t | sd t− 1sd− 1 t− 1 )

PθR-r,d (cηι, 1 | −, cη c′

ηι, −) def

= cηι = c′

ηι · EθG-rl∗

,d (cη

→cηι ...) EθG-rl∗

,d (cη ∗

→cηι ...)

PθR-r,d (cηι, 0 | −, cη c′

ηι, −) def

= cηι = c′

ηι · EθG-rl∗

,d (cη +

→cηι ...) EθG-rl∗

,d (cη ∗

→cηι ...)

William Schuler Incremental Parsing in Bounded Memory

slide-61
SLIDE 61

Preserving Dependencies in Sequence Model

Models now defined on CFG probs + chopped-up left-descendant terms... Transitions apply left/right CFG rules above/below incomplete constituent: PθS-t,d(sd

t | rd+ 1 t

rd

t sd t− 1sd− 1 t

) def = if rd

t =r⊥ : PθS-t-a,d (sd t | sd− 1 t

rd

t )

if rd

t =r⊥ : PθS-t-w,d(sd t | sd t− 1rd+ 1 t

) PθS-t-a,d(cηι, cηι1 | −, cη cηι0) def =

EθG-rl∗

,d (cη ∗

→cηι ...)·PθG-l,d (cηι→cηι0 cηι1) EθG-rl∗

,d (cη +

→cηι0 ...)

PθS-t-w,d(cη, cηι1 | c′

η, cηι cηι0) def

= cη = c′

η · PθG-r,d (cηι→cηι0 cηι1) EθG-rl∗

,d (cηι

→cηι0 ...)

William Schuler Incremental Parsing in Bounded Memory

slide-62
SLIDE 62

Preserving Dependencies in Sequence Model

Models now defined on CFG probs + chopped-up left-descendant terms... Transitions apply left/right CFG rules above/below incomplete constituent: PθS-t,d(sd

t | rd+ 1 t

rd

t sd t− 1sd− 1 t

) def = if rd

t =r⊥ : PθS-t-a,d (sd t | sd− 1 t

rd

t )

if rd

t =r⊥ : PθS-t-w,d(sd t | sd t− 1rd+ 1 t

) PθS-t-a,d(cηι, cηι1 | −, cη cηι0) def =

EθG-rl∗

,d (cη ∗

→cηι ...)·PθG-l,d (cηι→cηι0 cηι1) EθG-rl∗

,d (cη +

→cηι0 ...)

PθS-t-w,d(cη, cηι1 | c′

η, cηι cηι0) def

= cη = c′

η · PθG-r,d (cηι→cηι0 cηι1) EθG-rl∗

,d (cηι

→cηι0 ...)

Model thus preserves conditional dependencies of original PCFG. Same probabilities as CKY (mod. depth bounds): incrementalize any PCFG!

William Schuler Incremental Parsing in Bounded Memory

slide-63
SLIDE 63

Right-Corner Transform on Operator Chains

Preserving CFG dependencies preserves many predicate-argument relations. Allows incremental interpreter def. on bottom-up (compositional) semantics: NP{d2} NP{d1, d2, d3} Id directories Dir RR{c2} Mod0 VBG{c2, c3} Id containing Con NP{f1, f2} Arg1 executables Exe For example, interpretation may use probabilistic dependencies explicitly: PθG(lciη → lciη0 lciη1) = PθM(lcη0, lcη1 | lciη) · PθL(iη0 | lη0, iη) · PθL(iη1 | lη1, iη) (referent iη generates child relns, cats lcη0/1; relns generate referents iη0/1)

William Schuler Incremental Parsing in Bounded Memory

slide-64
SLIDE 64

Right-Corner Transform on Operator Chains

Transformed operations aligned to sequence model:

d=1 d=2 d=3 word t=1 t=2 t=3 directories/directories {d1d1, d2d2, d3d3} Dir −/− directories {d1, d2, d3} − NP/RR {d2c2, d3c3} Id,Mod0 containing/containing {c1c1, c2c2, c3c3} Id·Con NP/RR {d2c2, d3c3} − containing {c1, c2, c3} NP/executables {d2f2} Arg1·Exe −/− NP {d2} − −

William Schuler Incremental Parsing in Bounded Memory

slide-65
SLIDE 65

Evaluation: first-order denotations

Real-time speech interface (acoust model: Robinson’94, domain: 240 indiv.) Student domain: ‘go to sports track and set Homeroom 2 Bell to captain’

Homeroom2 T u r i n g Bell Sports F

  • t

b a l l Captain Defense Offense L i n e QuarterBack TightEnd T r a c k

e⊤ ehr2 eturing ebell esports efootball ecaptain edefense eoffense eline eqb ete etrack

William Schuler Incremental Parsing in Bounded Memory

slide-66
SLIDE 66

Evaluation: first-order denotations

Real-time speech interface (acoust model: Robinson’94, domain: 240 indiv.) Student domain: ‘go to sports track and set Homeroom 2 Bell to captain’

Homeroom2 T u r i n g Bell Sports F

  • t

b a l l Captain Defense Offense L i n e QuarterBack TightEnd T r a c k

e⊤ ehr2 eturing ebell esports efootball ecaptain edefense eoffense eline eqb ete etrack

Interactive model closer to human than trigram: concept error 26% → 17%

William Schuler Incremental Parsing in Bounded Memory

slide-67
SLIDE 67

Evaluation: second-order denotations

Bounded model allows 2ndorder denotations in real time speech (13% SER) Restaurant domain: ‘select the glasses on chairs’

William Schuler Incremental Parsing in Bounded Memory

slide-68
SLIDE 68

Non-local Dependencies

Model can also introduce non-CFG dependencies across store elements: Consider filler-gap construction: NP NP DT the NN bike S (gap:NP) NP you VP (gap:NP) VBD said S (gap:NP) NP DT your NN friend VP (gap:NP) VBD got PP IN for NP NNP christmas Conventional analysis requires feature passing [Pollard and Sag, 1994].

William Schuler Incremental Parsing in Bounded Memory

slide-69
SLIDE 69

Non-local Dependencies

In sequence model, referent is locally available at gap position:

d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 t=7 t=8 t=9 t h e b i k e y

  • u

s a i d y

  • u

r f r i e n d g

  • t

f

  • r

c h r i s t m a s − − − − − N P / N N − − − − − − S x / V P x S x / S x S x / S x S x / V P x S x / P P S x / N P − N P / N N N P / G C n p S N P / G C n p S N P / G C n p S N P / G C n p S N P / G C n p S N P / G C n p S N P / G C n p S

Still need ‘x’ feature to require gap, but no need to pass referent/category (obtained from tree by restricting awaited transitions on GC category)

William Schuler Incremental Parsing in Bounded Memory

slide-70
SLIDE 70

Non-local Dependencies

Crossed and nested dependencies respect chunks, but require some relaxation in definition of reduction:

d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 J a n P i e t M a r i e z a g l a t e n z w e m m e n − − − S / V P S / V P S / V P − − S / V P S / V P S / V P S / S − S / V P S / V P S / V P S / S S / S

(no longer easily associated with phrase structure tree)

William Schuler Incremental Parsing in Bounded Memory

slide-71
SLIDE 71

Non-local Dependencies

Crossed and nested dependencies respect chunks, but require some relaxation in definition of reduction:

d=1 d=2 d=3 word t=1 t=2 t=3 t=4 t=5 t=6 J a n P i e t M a r i e z a g l a t e n z w e m m e n − − − S / V P S / V P S / V P − − S / V P S / V P S / V P S / S − S / V P S / V P S / V P S / S S / S

(no longer easily associated with phrase structure tree) Speculate EPDA-style transition operations could be learned, but PDA-style (HHMM) operations usually minimize surprisal.

William Schuler Incremental Parsing in Bounded Memory

slide-72
SLIDE 72

Conclusion

Simple processing model, matches observations about human memory

  • 1. bounded number of unconnected chunks

[Miller, 1956, Cowan, 2001] (subjects group stimuli into only 4 or so clusters)

  • 2. process rich syntax

[Chomsky and Miller, 1963] (center embedding: ‘if [neither [the man [the cop] saw] nor ...] then ...’) (cf. center recursion: ‘?? the malt [the rat [the cat chased] ate] ...’)

  • 3. processing is incremental

[Sachs, 1967, Jarvella, 1971] (subjects can’t remember specific words earlier in sentence)

  • 4. processing is parallel, probabilistic

[Jurafsky, 1996, Hale, 2001, Levy, 2008] (probabilistic / info. theoretic measures correlate w. reading times)

William Schuler Incremental Parsing in Bounded Memory

slide-73
SLIDE 73

Conclusion

Simple processing model, matches observations about human memory In particular, the factored sequence model (dynamic Bayes net):

  • 1. has easily interpretable random variables

(explicit estimation of speaker intent; cf. neural net)

  • 2. has clear role of bounded working memory store

(random variable for each store element)

  • 3. has clear role of syntax

(grammar transform turns trees into chunks for store elements)

  • 4. is fast enough to interact with real-time speech recognizer

(using cool engineering tricks: best-first / ‘lazy’ k-best search) Result is a nice platform for linguistic experimentation!

William Schuler Incremental Parsing in Bounded Memory

slide-74
SLIDE 74

Bibliography I

Abney, S. P. and Johnson, M. (1991). Memory requirements and local ambiguities of parsing strategies.

  • J. Psycholinguistic Research, 20(3):233–250.

Bachrach, A., Roark, B., Marantz, A., Whitfield-Gabrieli, S., Cardenas, C., and Gabrieli, J. D. (2009). Incremental prediction in naturalistic language processing: An fMRI study. Bellman, R. (1957). Dynamic Programming. Princeton University Press, Princeton, NJ. Chomsky, N. and Miller, G. A. (1963). Introduction to the formal analysis of natural languages. In Handbook of Mathematical Psychology, pages 269–321. Wiley.

William Schuler Incremental Parsing in Bounded Memory

slide-75
SLIDE 75

Bibliography II

Cowan, N. (2001). The magical number 4 in short-term memory: A reconsideration of mental storage capacity. Behavioral and Brain Sciences, 24:87–185. Hale, J. (2001). A probabilistic earley parser as a psycholinguistic model. In Proceedings of the Second Meeting of the North American Chapter

  • f the Association for Computational Linguistics, pages 159–166,

Pittsburgh, PA. Jarvella, R. J. (1971). Syntactic processing of connected speech. Journal of Verbal Learning and Verbal Behavior, 10:409–416.

William Schuler Incremental Parsing in Bounded Memory

slide-76
SLIDE 76

Bibliography III

Johnson, M. (1998). Finite state approximation of constraint-based grammars using left-corner grammar transforms. In Proceedings of COLING/ACL, pages 619–623, Montreal, Canada. Jurafsky, D. (1996). A probabilistic model of lexical and syntactic access and disambiguation. Cognitive Science: A Multidisciplinary Journal, 20(2):137–194. Klein, D. and Manning, C. D. (2003). Accurate unlexicalized parsing. In Proceedings of the 41st Annual Meeting of the Association for Computational Linguistics, pages 423–430, Sapporo, Japan.

William Schuler Incremental Parsing in Bounded Memory

slide-77
SLIDE 77

Bibliography IV

Levy, R. (2008). Expectation-based syntactic comprehension. Cognition, 106(3):1126–1177. Miller, G. A. (1956). The magical number seven, plus or minus two: Some limits on our capacity for processing information. Psychological Review, 63:81–97. Miller, T. (2009). Improved syntactic models for parsing speech with repairs. In Proceedings of the North American Association for Computational Linguistics, Boulder, CO.

William Schuler Incremental Parsing in Bounded Memory

slide-78
SLIDE 78

Bibliography V

Miller, T. and Schuler, W. (2008). A syntactic time-series model for parsing fluent and disfluent speech. In Proceedings of the 22nd International Conference on Computational Linguistics (COLING’08). Miller, T. and Schuler, W. (2010). Hhmm parsing with limited parallelism. In Proceedings of the Workshop on Cognitive Modeling and Computational Linguistics ’10). Pollard, C. and Sag, I. (1994). Head-driven Phrase Structure Grammar. University of Chicago Press, Chicago.

William Schuler Incremental Parsing in Bounded Memory

slide-79
SLIDE 79

Bibliography VI

Roark, B., Bachrach, A., Cardenas, C., and Pallier, C. (2009). Deriving lexical and syntactic expectation-based measures for psycholinguistic modeling via incremental top-down parsing. Proceedings of the 2009 Conference on Empirical Methods in Natural Langauge Processing, pages 324–333. Sachs, J. (1967). Recognition memory for syntactic and semantic aspects of connected discourse. Perception and Psychophysics, 2:437–442. Schuler, W., AbdelRahman, S., Miller, T., and Schwartz, L. (2008). Toward a psycholinguistically-motivated model of language. In Proceedings of COLING, pages 785–792, Manchester, UK.

William Schuler Incremental Parsing in Bounded Memory

slide-80
SLIDE 80

Bibliography VII

Schuler, W., AbdelRahman, S., Miller, T., and Schwartz, L. (2010). Broad-coverage incremental parsing using human-like memory constraints. Computational Linguistics, 36(1). Steedman, M. (2000). The syntactic process. MIT Press/Bradford Books, Cambridge, MA. Wu, S., Bachrach, A., Cardenas, C., and Schuler, W. (2010). Complexity metrics in an incremental right-corner parser. In Proceedings of the 49th Annual Conference of the Association for Computational Linguistics (ACL’10).

William Schuler Incremental Parsing in Bounded Memory