SemLink+: FrameNet, VerbNet, and Event Ontologies Martha Palmer, - - PowerPoint PPT Presentation

semlink framenet verbnet and event ontologies
SMART_READER_LITE
LIVE PREVIEW

SemLink+: FrameNet, VerbNet, and Event Ontologies Martha Palmer, - - PowerPoint PPT Presentation

SemLink+: FrameNet, VerbNet, and Event Ontologies Martha Palmer, Claire Bonial, Diana McCarthy University of Colorado Frame Semantics in NLP: A Workshop in Honor of Chuck Fillmore (1929 2014) ACL Workshop June 27, 2014 1 Outline n


slide-1
SLIDE 1

SemLink+: FrameNet, VerbNet, and Event Ontologies

Martha Palmer, Claire Bonial, Diana McCarthy University of Colorado Frame Semantics in NLP: A Workshop in Honor of Chuck Fillmore (1929 – 2014) ACL Workshop June 27, 2014

1

slide-2
SLIDE 2

Outline

n Deep NLU? n Where we are now n Where we need to go n More details about where we need to go n The contributions and limitations of lexical

resources to this process

2

slide-3
SLIDE 3

Where we are now – shallow semantics

n Syntactic Structure – parse trees, Treebanks n Semantic types – nominal entities [Person,

Location, Organization], NE tagging

n Semantic roles – Agents, [PropBank FrameNet,

VerbNet]

n Sense distinctions – call me a taxi, call me an

idiot, WordNet, OntoNotes groups, FrameNet, VerbNet, vectors, etc.

n Coreference – [President Obama: he]

3

slide-4
SLIDE 4

Where we are now - DETAILS

n DARPA-GALE, OntoNotes 5.0

q BBN, Brandeis, Colorado, Penn q Multilayer structure: NE, TB, PB, WS, Coref q Three languages: English, Arabic, Chinese q Several Genres (@ ≥ 200K ): NW, BN, BC, WT

n Close to 2M words @ language (less PB for Arabic)

q Parallel data, E/C, E/A

n DARPA BOLT – discussion forum, SMS

q PropBank extensions: light verbs, function tags

  • n core args, nominalizations, adjectives,

constructions, often relying on FrameNet

4

slide-5
SLIDE 5

PropBank Verb Frames Coverage

94% 95% 96% 97% 98% 99% 100% 1000 2000 3000 4000 5000 6000 7000 8000

n The set of verbs is open n But the distribution is

highly skewed

n For English, the 1000

most frequent lemmas cover 95% of the verbs in running text.

q

Graphs show counts over English Web data containing 150 M verbs. 5

FrameNet and VerbNet should have the same coverage, and we (or at least VerbNet) desperately need help to do this semi-automatically!!

slide-6
SLIDE 6

WordNet: - call, 28 senses, 9 groups

WN5, WN16,WN12 WN15 WN26 WN3 WN19 WN4 WN 7 WN8 WN9 WN1 WN22 WN20 WN25 WN18 WN27 WN2 WN 13 WN6 WN23 WN28 WN17 , WN 11 WN10, WN14, WN21, WN24, Loud cry Label Phone/radio Bird or animal cry Request Call a loan/bond Visit Challenge Bid

slide-7
SLIDE 7

SEMLINK-PropBank, VerbNet, FrameNet, WordNet, OntoNotes

fit-54.3, ON3

WN1 WN2 WN5 WN20 WN22 WN24 WN24 WN31 WN33 WN34 WN1 WN3 WN8 WN11 WN 23 WN9 WN16 WN17 WN19 WN27 WN37 WN38 WN28 WN32 WN35 WN36 ON4 – win election PropBank Frameset1*

carry

Palmer, Dang & Fellbaum, NLE 2007

carry-11.4, CARRY,-FN ,ON1 cost-54.2, ON2

*ON5-ON11 carry oneself,carried away/out/off, carry to term

slide-8
SLIDE 8

8

Sense Hierarchy

n PropBank Framesets – ITA >90%

coarse grained distinctions 20 Senseval2 verbs w/ > 1 Frameset Maxent WSD system, 73.5% baseline, 90%

q Sense Groups (Senseval-2/OntoNotes) - ITA 89%

Intermediate level (includes Verbnet/some FrameNet) – SVM, 88+%

n WordNet – ITA 73%

fine grained distinctions, 64%

Dligach & Palmer, ACL2011

slide-9
SLIDE 9

9

SEMLINK

n Extended VerbNet: 6,340 senses

n 92% PB tokens (8114 verb senses/12,646 all)

n Type-type mapping PB/VN, VN/FN, VN/WN n Semi-automatic mapping of WSJ PropBank

instances to VerbNet classes and thematic roles, hand-corrected. (now FrameNet also)

n VerbNet class tagging as automatic WSD n Run SRL, map Arg2 to VerbNet roles, Brown

performance improves

Yi, Loper, Palmer, NAACL07 Brown, Dligach, Palmer, IWCS 2011; Croce, et. al., ACL2012

slide-10
SLIDE 10

Where we need to go – Richer Event Descriptions - RED

n “Saucedo said that guerrillas in one car opened

fire on police standing guard, while a second car carrying 88 pounds (40 kgs) of dynamite parked in front of the building, and a third car rushed the attackers away.”

n Saucedo said – reporting event, evidential

10

slide-11
SLIDE 11

What we can do

n that guerrillas in one car opened fire on police

standing guard

n opened fire = aspectual context,

q fire(guerillas, police)

n standing guard = support verb construction/

aspectual?, reduced relative

q guard(police, X)

11

slide-12
SLIDE 12

What we can do, cont.

n while a second car carrying 88 pounds (40 kgs) of

dynamite parked in front of the building

n carrying - reduced relative, correct head

noun - pounds or dynamite?

q carry(car2, dynamite)

n park(car2, front_of(building))

12

slide-13
SLIDE 13

What we can do, cont.

n and a third car rushed the attackers away n rush(car3, attackers, away)

13

slide-14
SLIDE 14

Temporal & Causal ordering?

n “Saucedo said that guerrillas in one car opened

fire on police standing guard, while a second car carrying 88 pounds (40 kgs) of dynamite parked in front of the building, and a third car rushed the attackers away

n guarding BEFORE/OVERLAP firing n Narrative container – TimeX

q [firing, parking, rushing] all overlap, all in the

same temporal bucket?

q [see Styler, et. al, ACL2014, Events Workshop & RED

Guidelines]

14

slide-15
SLIDE 15

Don’t mark the relations between EVENT s.

!

Instead, put EVENT s in temporal buckets and relate the buckets

slide-16
SLIDE 16
slide-17
SLIDE 17

Temporal & Causal ordering

n “Saucedo said that guerrillas in one car opened

fire on police standing guard, while a second car carrying 88 pounds (40 kgs) of dynamite parked in front of the building, and a third car rushed the attackers away

q guarding BEFORE/OVERLAP firing q X CONTAINS [firing, parking, rushing] q firing BEFORE parking q parking BEFORE rushed

17

slide-18
SLIDE 18

Implicit arguments

n that guerrillas in one car opened fire on police

standing guard

n opened fire = aspectual context,

q fire(guerillas, police)

n standing guard = support verb construction or

aspectual?, reduced relative

q guard(police, X)

18

slide-19
SLIDE 19

More compelling example

(thanks to Vivek Srikumar)

n The bomb exploded in a crowded marketplace.

Five civilians were killed, including two children. Al Qaeda claimed responsibility.

n Killed by Whom? n Responsibility for what? n Need recovery of implicit arguments

19

slide-20
SLIDE 20

VerbNet – based on Levin, B.,93 Kipper, et. al., LRE08

n Class entries:

q Capture generalizations about verb behavior q Organized hierarchically q Members have common semantic elements,

semantic roles, syntactic frames, predicates

n Verb entries:

q Refer to a set of classes (different senses) q each class member linked to WN synset(s), ON

groupings, PB frame files, FrameNet frames,

20

slide-21
SLIDE 21

VerbNet: send-11.1 (Members: 11, Frames: 5)

includes “ship”

n Roles

n Agent [+animate | +organization] n Theme [+concrete] n Source [+location] n Destination [+animate | [+location & -region]]

n Syntactic Frame:NP V NP PP.destination

q example

"Nora sent the book to London."

q syntax

Agent V Theme {to} Destination

q semantics motion(during(E), Theme)

location(end(E), Theme, Destination) cause(Agent, E)

21

slide-22
SLIDE 22

22

Recovering Implicit Arguments*

[Palmer, et. al., 1986; Gerber & Chai, 2010, 2012; Ruppenhofer, Sporleder, Morante, Baker, Palmer. 2010. SemevEval-2010 Task10:]

[Arg0 The two companies] [REL1 produce] [Arg1 market pulp, containerboard and white paper]. The goods could be manufactured closer to customers, saving [REL2 shipping] costs.

n Used VerbNet for subcategorization frames

* AKA definite null complements

slide-23
SLIDE 23

Implicit arguments

n SYNTAX

Agent V Theme {to} Destination [AGENT] shipped [THEME] to [DESTINATION]

n SEMANTICS

q CAUSE(AGENT,E) q MOTION(DURING(E), THEME), q LOCATION(END(E), THEME, DESTINATION),

23

slide-24
SLIDE 24

Implicit arguments instantiated using coreference

n [AGENT] shipped [THEME] to [DESTINATION] n [Companies] shipped [goods] to [customers]. n SEMANTICS

q CAUSE(Companies, E) q MOTION(DURING(E), goods), q LOCATION(END(E), goods, customers),

24

Can annotate, semi-automatically!

slide-25
SLIDE 25

Another type of Implicit Relation

Example from Daniel Marcu, GALE Wrap-up Mtg

n Between Munich and LA you need less than 11

hours by plane.

n You can fly

to Los Angeles from Munchen in no more than eleven hours.

n From Munich to Los Angeles,

it does not take more than eleven hours by plane.

slide-26
SLIDE 26

Constructions allow us to

n Recognize a path prepositional phrase, and

that it necessarily goes with a “MOTION” event – Caused-motion constructions

q John sneezed the tissue off the table. q Mary blinked the snow off of her eyelashes.

n If we detect a MOTION event we can

associate the plane with it as a vehicle

n Just the plane itself can suggest a motion

event…

26

slide-27
SLIDE 27

Construction Grammar

n In Construction Grammar

(Fillmore, 1988, Goldberg, 1995, Kay and Fillmore, 1999, Michaelis, 2004, Goldberg, 2005)

q constructions are carriers of meaning q constructions are assigned meaning in the same

way that words are – via convention rather than composition.

n Invaluable resource –

FrameNet Constructicon, Cxn Viewer

27

slide-28
SLIDE 28

2 8

n Introduce a constructional ``layer” to VerbNet,

which attaches orthogonally to relevant VerbNet classes

Caused Motion Resultative

Constructional Layer

manner_speaking-37.3 hiccup-40.1.1 weather-57

Current VN

They hissed him out of the university. He blinked the snow off his eyelashes. He blinked his eyes dry. The pond froze solid.

Introducing a Constructional Layer to VerbNet

47

Jena Hwang, LREC-2014

slide-29
SLIDE 29

29

VerbNet can also provide

inferences – sometimes…

q Every path from back door to yard was covered by

a grape-arbor, and every yard had fruit trees.

q Where are the grape arbors located?

slide-30
SLIDE 30

30

VerbNet – cover, fill-9.8 class

n Members: fill, …, cover,…, staff, …. n Thematic Roles: Agent

Theme Destination

n Syntactic Frames with Semantic Roles

“The employees staffed the store" “ The grape arbors covered every path"

Theme V Destination location(E,Theme,Destination) location(E,grape_arbor,path)

slide-31
SLIDE 31

Inferences can inform Coreference?

n “Saucedo said that guerrillas in one car opened

fire on police standing guard, while a second car carrying 88 pounds (40 kgs) of dynamite parked in front of the building, and a third car rushed the attackers away.”

n AMR

q guerilla – Arg0 of fire.01 q attacker – person – Arg0-of attack.01

31

slide-32
SLIDE 32

FrameNet Attack frame

32

attacker.n, fire.n Loosely connected to “fire.v” via Hostile Encounter

slide-33
SLIDE 33

Needed – An Event Ontology

n That can provide appropriate levels of

generalization

n DEFT - Event Ontology conference calls

q Martha Palmer, James Pustejovsky, Annie Zaenen,

Diana McCarthy, Teruko Mitamura, German Rigau, Ann Bies, Kira Griffit, Julie Fitzgerald, Claire Bonial, Derek Palmer

n Map ERE event types to FrameNet? n Develop an upper level event ontology that

ERE and FN can both be mapped to?

33

slide-34
SLIDE 34

Conflict events

n ERE

n Attack events n Protest/Demonstration events

n FrameNet

n Attack events - See previous slide n Protest, not present n Demonstration – Reasoning frame

n VerbNet

n Attack - Judgment n Protest - Conspire n Demonstrate – Transfer_message

34

slide-35
SLIDE 35

Eventuality ¡ FN: ¡Protest ¡ (demonstrate, ¡protest) ¡ ERE: ¡Conflict.demonstrate ¡ A?ack ¡ FN: ¡A?ack ¡ (assail, ¡hit, ¡bomb) ¡ State ¡ Ba?le ¡ (ba?le, ¡clash, ¡duel) ¡ ¡ Social ¡InteracFon ¡ FN: ¡IntenFonally_act ¡ (do, ¡act, ¡engage) ¡ Social ¡Encounter ¡

VN: ¡Conspire-­‑71 ¡ VN: ¡TBD ¡ VN: ¡Ba?le-­‑36.4 ¡

ERE: ¡Conflict ¡ ¡Accord ¡ Judgment ¡ FN: ¡Taking_sides ¡ (endorse, ¡oppose) ¡ OpposiFon ¡ PHYSICAL ¡CompeFFon ¡ FN: ¡HosFle_encounter ¡ ERE: ¡Conflict.a?ack ¡ NON-­‑PHYSICAL ¡ CompeFFon ¡

PRECEDES ¡

slide-36
SLIDE 36
slide-37
SLIDE 37
slide-38
SLIDE 38

Using ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡for ¡Ontologies ¡

  • Will ¡coordinate ¡with ¡SUMO, ¡WN ¡also ¡
  • Any ¡advice? ¡
  • Could ¡really ¡use ¡some ¡help ¡from ¡an ¡

experienced ¡user ¡

  • Could ¡also ¡REALLY ¡use ¡input ¡from ¡all ¡of ¡these ¡

“clustering” ¡techniques! ¡

38

slide-39
SLIDE 39

6

Predicate ¡Matrix ¡- Lacalle, Laparra, Rigau, LREC 2014 -

slide-40
SLIDE 40

7

Predicate Matrix

slide-41
SLIDE 41

41

Predicate ¡Matrix ¡

First version 1.0 (GWC 2014)

SemLink +

Monosemous verbs from VN + Synonyms from WN

  • Second version 1.1 (LREC 2014)

SemLink +

Automatic mappings between predicates +

  • WN-VN and WN-FN (new mappings!)

Project VN roles to FN roles (complete gaps!) +

Synonyms from WN

slide-42
SLIDE 42

Where we need to go

n Recovery of implicit arguments n Recovery of implicit relations n Better Entity coreference n Event coreference n Temporal and Causal ordering of events n Generalizations over event types

42

slide-43
SLIDE 43

Lexical resources can provide

n Generalizations about subcat frames & roles n Backoff classes for OOV items for portability n Semantic similarities/”types” for verbs n Event type hierarchies for inferencing n Need to be unified and empirically validated and

extended: Semlink+

q VN & FN need PB like coverage, and techniques for

extension and automatic domain adaptation

n Hybrid lexicons – symbolic and statistical

lexical entries?

43

slide-44
SLIDE 44

Acknowledgments

n We gratefully acknowledge the support of the National

Science Foundation Grants for Consistent Criteria for Word Sense Disambiguation, Robust Semantic Parsing, Richer Representations for Machine Translation, A Bayesian Approach to Dynamic Lexical Resources for Flexible Language Processing, DARPA-GALE via a subcontract from BBN, DARPA-BOLT & DEFT via a subcontract from LDC, and DTRA: SemLink+ via a subcontract from BBN, and NIH THYME.

n Any opinions, findings, and conclusions or

recommendations expressed in this material are those of the authors and do not necessarily reflect the views of the National Science Foundation, DARPA or NIH.

44

slide-45
SLIDE 45

And thanks to

n Postdocs: Paul Kingsbury, Dan Gildea,

Nianwen Xue, Jinying Chen

n Students: Joseph Rosenzweig, Hoa Dang,

Tom Morton, Karin Kipper Schuler, Jinying Chen, Szu-Ting Yi, Edward Loper, Susan Brown, Dmitriy Dligach, Jena Hwang, Will Corvey, Claire Bonial, Jinho Choi, Lee Becker, Shumin Wu, Kevin Stowe

n Collaborators: Christiane Fellbaum, Suzanne

Stevenson, Annie Zaenen, Orin Hargraves, James Pustejovsky

45