Surface Construction Labeling Lori Levin Language Technologies - - PowerPoint PPT Presentation

surface construction labeling
SMART_READER_LITE
LIVE PREVIEW

Surface Construction Labeling Lori Levin Language Technologies - - PowerPoint PPT Presentation

Corpus Annotation for Surface Construction Labeling Lori Levin Language Technologies Institute Carnegie Mellon University Collaborators and acknowledgements Collaborators: Jesse Dunietz, Jaime Carbonell, Dunietz, Jesse, Lori Levin, and


slide-1
SLIDE 1

Corpus Annotation for Surface Construction Labeling

Lori Levin Language Technologies Institute Carnegie Mellon University

slide-2
SLIDE 2

Collaborators and acknowledgements

  • Collaborators: Jesse Dunietz, Jaime Carbonell,

– Dunietz, Jesse, Lori Levin, and Jaime Carbonell. "Automatically Tagging Constructions of Causation and Their Slot-Fillers." Transactions of the Association for Computational Linguistics 5 (2017). 117-133 – Dunietz, Jesse, Lori Levin, and Jaime Carbonell. "The BECauSE Corpus 2.0: Annotating Causality and Overlapping Relations." Proceedings of LAW XI – The 11th Linguistic Annotation Workshop (2017). 95-104. – Dunietz, Jesse, Lori Levin, and Jaime Carbonell. "Annotating Causal Language Using Corpus Lexicography of Constructions." Proceedings of LAW IX – The 9th Linguistic Annotation Workshop(2015): 188-196.

– Dunietz, Jesse, Lori Levin, and Miriam R. L. Petruck. "Construction Detection in a Conventional NLP Pipeline." AAAI Spring Symposium Technical Report SS-17-02: Computational Construction Grammar and Natural Language Understanding (2017).

– Jesse Dunietz, Annotating and Automatically Tagging Constructions of Causal Language, Ph.D. Thesis, Carnegie Mellon University, 2018.

– Jesse Dunietz, Jaime Carbonell and Lori Levin, “DeepCx: A transition-based approach for shallow semantic parsing with complex constructional triggers”. EMNLP 2018

  • Thank you: Nathan Schneider, Miriam Petruck, Alexis Palmer
slide-3
SLIDE 3

Outline

  • What is Surface Construction Labeling (SCL)?
  • The causality task (Dunietz 2018)
  • The Because onomasiological annotation scheme
  • Making the task doable

– Constructicon as an annotation tool – Overlapping Relations

  • Slow research musings about grammaticalization

and the contents of the constructicon

slide-4
SLIDE 4

Surface Construction Labeling is a type of Frame Semantic Parsing (aka Shallow Semantic Parsing, Semantic Role Labeling)

http://www.cs.cmu.edu/~ark/SEMAFOR/

slide-5
SLIDE 5

Surface Construction Labeling is a type of Frame Semantic Parsing (aka Shallow Semantic Parsing, Semantic Role Labeling)

http://www.cs.cmu.edu/~ark/SEMAFOR/

Names of semantic Frames from the FrameNet lexicon

slide-6
SLIDE 6

Surface Construction Labeling is a type of Frame Semantic Parsing (aka Shallow Semantic Parsing, Semantic Role Labeling)

http://www.cs.cmu.edu/~ark/SEMAFOR/

Names of semantic Frames from the FrameNet lexicon Frame-Evoking Elements

slide-7
SLIDE 7

Surface Construction Labeling is a type of Frame Semantic Parsing (aka Shallow Semantic Parsing, Semantic Role Labeling)

http://www.cs.cmu.edu/~ark/SEMAFOR/

Names of semantic Frames from the FrameNet lexicon Frame-Evoking Elements Each Frame-Evoking Element is a lexical item

  • r MWE.
slide-8
SLIDE 8

Surface Construction Labeling is a type of Frame Semantic Parsing (aka Shallow Semantic Parsing, Semantic Role Labeling)

http://www.cs.cmu.edu/~ark/SEMAFOR/

Frame Elements of the DESIRING frame: Experiencer and Event Each Frame Element has a label and an extent.

slide-9
SLIDE 9

Example of Surface Construction Labeling (in a notation that is easier to format) She got so lonely that she decided to watch TV Cause Effect Motivation (enable)

Frame Name and Frame Evoking Elements

slide-10
SLIDE 10

Example of Surface Construction Labeling (in a notation that is easier to format) She got so lonely that she decided to watch TV Cause Effect Motivation (enable)

Frame Name and Frame Evoking Elements The frame name is from the annotation scheme that I will be describing in this talk. The frame indicates that being excessively lonely was her motivation for watching TV.

slide-11
SLIDE 11

Example of Surface Construction Labeling (in a notation that is easier to format) She got so lonely that she decided to watch TV Cause Effect Motivation (enable)

Frame Name and Frame Evoking Elements The Frame-Evoking Element(s) is not a single lexical item. It is a construction consisting of an adjective phrase with an intensifier and a finite clausal complement.

slide-12
SLIDE 12

Example of Surface Construction Labeling

She got so lonely that she decided to watch TV Cause Effect Motivation (enable)

Constructions: (Fillmore, Kay, and O’Connor, 1988) pairings of form and meaning, where the form may be any constellation of morpho-syntactic or lexical material. A Constructicon is repository of constructions where the Frame Evoking Elements can an arbitrary constellation of forms and the meaning side evokes a frame. (Fillmore et al., 2011)

slide-13
SLIDE 13
  • Expand the scope of frame semantic parsing
  • By allowing Frame Evoking Elements to be constructs,

arbitrary constellations of morpho-syntactic and lexical elements.

  • Creating a unified approach to Frame Semantic Parsing,

integrating the FrameNet lexicon and the FrameNet Constructicon

  • Resulting in broader coverage of more ways to express

each frame

Goal of Surface Construction Labeling

slide-14
SLIDE 14

Why is it called Surface Construction Labeling? She got so lonely that she decided to watch TV.

Full construction labeling (aka “constructions all the way down”) would parse all of the constructions in the sentence and their composition into larger constructions.

  • Definite referring expression

construction, intensified adjective construction, etc. SCL recognizes constructions shallowly

  • n top of Stanford NLP parses.
slide-15
SLIDE 15

Why is it called Surface Construction Labeling? She got so lonely that she decided to watch TV.

Full construction labeling (aka “constructions all the way down”) would parse all of the constructions in the sentence and their composition into larger constructions.

  • Definite referring expression

construction, intensified adjective construction, etc. SCL recognizes constructions shallowly

  • n top of Stanford NLP parses.
slide-16
SLIDE 16

Outline

  • What is Surface Construction Labeling (SCL)?
  • The causality task (Dunietz 2018)
  • The Because onomasiological annotation scheme
  • Making the task doable

– Constructicon as an annotation tool – Overlapping Relations

  • Slow research musings about grammaticalization

and the contents of the constructicon

slide-17
SLIDE 17

CAUSE CAUSATION

(ENABLEMENT)

The hot summer

EFFECT

set the stage for the devastating fires . Why

CAUSATION

Annotating Causality

is California on fire ?

EFFECT

slide-18
SLIDE 18

CAUSE CAUSATION

(ENABLEMENT)

The hot summer

EFFECT

set the stage for the devastating fires . Why

CAUSATION

Annotating Causality

is California on fire ?

EFFECT

Connective Argument spans

slide-19
SLIDE 19

Causality is way too hard. Why did we choose it?

slide-20
SLIDE 20

Variety in Frame Evoking Elements

Such swelling can impede breathing. They moved because of the schools. We’re running late, so let’s move quickly. Our success is contingent on your support. This opens the way for broader regulation. Judy’s comments were so offensive that I left. For markets to work, banks can’t expect bailouts. (Verbs) (Prepositions) (Conjunctions) (Adjectives) (Multi-word expr.s) (Complex) (Complex)

slide-21
SLIDE 21

Variety in syntactic positions of FEs with respect to FEEs

For nominal FEEs, the EFFECT may not be a syntactic complement of the FEE The cause of her illness was dehydration. Her illness’ cause was dehydration. Her chart listed her illness’ cause as dehydration.

slide-22
SLIDE 22

33% of explicit relations between French verbs (Conrath et al. 2011) 12% of explicit discourse connectives in Penn Discourse Treebank (Prasad et al., 2008)

Causality is not rare

>5% and among the most complex

  • f questions asked to question-answering systems

(Verberne et al., 2010)

slide-23
SLIDE 23

23

(Temporal) (Extremity) (Correlation) (Permission) (Temporal + (Correlation) After a drink, she felt much better. They’re too big to fail. The more I read his work, the less I like it. The police let his sister visit him briefly. As voters get to know Mr. Romney, his poll numbers will rise.

Causal language is difficult to disentangle from overlapping semantic domains.

slide-24
SLIDE 24

85.2% 62.8% 63.6% 63.4% 65.4% 18.1% 46.2% 44.6% 55.8% 56.5% 29.5% 53.1% 52.3% 59.2% 60.5% Benchmark Benchmark + CW-S Benchmark + CW-L DeepCx DeepCx (known conn.s) Precision Recall F1

Automatic labeling of FEEs

Neural net won

slide-25
SLIDE 25

Automatic labeling of FEs

22.3% 37.9% 38.5% 43.9% 44.7% 73.5% 24.7% 42.5% 43.5% 50.6% 51.0% 80.9% 15.4% 24.8% 30.6% 41.0% 42.8% 67.8% 22.4% 38.7% 40.7% 51.7% 52.6% 82.8% Benchmark Benchmark + CW-S Benchmark + CW-L DeepCx DeepCx (known conn.s) DeepCx (oracle conn.s) Cause (exact) Cause (50%) Effect (exact) Effect (50%) Neural net won

slide-26
SLIDE 26

FrameNet

(Fillmore & Baker, 2010; Ruppenhofer et al., 2016)

Penn Discourse Treebank

(Prasad et al., 2008)

Each existing semantic parsing representation handles only a portion of this space.

CONTINGENCY:Cause ARG1

so Its products are simpler ,

ARG 2

customers need less assistance .

Conjunctions & adverbs only

(one) meaning

  • ne

word

CAUSATION

EFFECT

EFFECT

CAUSER

made bow me He

PURPO

SE

to show his dominance Words or constituents

  • nly

PropBank

(Palmer et al., 2005)

MAKE.02 ARG2 ARG1 ARG0

made bow me He

ARGM-PRP

to show his dominance . Verbs only

slide-27
SLIDE 27

We did some analysis, and the neural net really won. So, unfortunately, that has to be presented in EMNLP. Jesse Dunietz, Jaime Carbonell and Lori Levin, “DeepCx: A transition-based approach for shallow semantic parsing with complex constructional triggers”. EMNLP 2018

slide-28
SLIDE 28

Outline

  • What is Surface Construction Labeling (SCL)?
  • The causality task (Dunietz 2018)
  • The Because onomasiological annotation scheme
  • Making the task doable

– Constructicon as an annotation tool – Overlapping Relations

  • Slow research musings about grammaticalization

and the contents of the constructicon

slide-29
SLIDE 29

Annotation for Surface Construction Labeling

  • Onomasiological

– “How do you express x?”

  • https://en.wikipedia.org/wiki/Onomasiology
  • Hasegawa et al.
  • For any potential annotation unit:

– Is meaning X expressed? (semasiological) – Is there a form that expresses X?

  • If you do this as a corpus study, you end up with a

collection of constructions that express X in the corpus.

slide-30
SLIDE 30

The Because Corpus

Documents

Sentences Causal

New York Times Washington section

(Sandhaus, 2014)

59 1924 717 Penn TreeBank WSJ 47 1542 534 2014 NLP Unshared Task in PoliInformatics

(Smith et al., 2014)

3 772 324 Manually Annotated Sub- Corpus

(Ide et al., 2010)

12 629 228 Total 121 4790 1803

BECAUSE = Bank of Effects and Causes Stated Explicitly

slide-31
SLIDE 31

The Because Annotation Scheme

  • Bank of Effects and Causes Stated Explicitly
slide-32
SLIDE 32

Causal language: a clause or phrase in which

  • ne event, state, action, or entity

is explicitly presented as promoting or hindering another

slide-33
SLIDE 33

Causal language: a clause or phrase in which

  • ne event, state, action, or entity

is explicitly presented as promoting or hindering another

Annotators can’t do this.

slide-34
SLIDE 34

We annotate three types of causation.

The system failed because of a loose screw.

CONSEQUENCE

Mary left because John was there.

MOTIVATION

Mary left in order to avoid John.

PURPOSE

For us to succeed, we all have to cooperate We all have to cooperate (cause) for us to succeed (effect)

slide-35
SLIDE 35

Motivation and Purpose

  • Motivation: the reason exists

– Mary left because John was there

  • John being there (cause) motivated Mary to leave

(effect)

  • Purpose: a desired state

– Mary left in order to avoid John – (Mary wants to avoid John) causes (Mary leaves) – (Mary leaves) may cause/enable (Mary avoids John)

For us to succeed, we all have to cooperate We all have to cooperate (cause) for us to succeed (effect)

slide-36
SLIDE 36

Causation can be positive or negative.

This has often caused problems elsewhere.

FACILITATE

He kept the dog from leaping at her.

INHIBIT

slide-37
SLIDE 37

Connective: fixed morphological or

lexical cue indicating a causal construction

John killed the dog because it was threatening his chickens. John prevented the dog from eating his chickens. Ice cream consumption causes drowning.

We do not annotate most agentive verbs. We

  • nly annotate verbs that express causation as

their primary meaning.

slide-38
SLIDE 38

Effect: presented as outcome Cause: presented as producing effect

John killed the dog because it was threatening his chickens. John prevented the dog from eating his chickens. Ice cream consumption causes drowning. She must have met him before, because she recognized him yesterday.

slide-39
SLIDE 39

Means arguments for cases with an agent and an action

caused a commotion My dad by shattering a glass .

MEANS EFFECT CAUSE

By altering immune responses, inflammation can trigger depression.

We have not yet added the AFFECTED frame element (Rehbein and Ruppenhofer 2017).

slide-40
SLIDE 40

We exclude language that does not encode pure, explicit causation:

Relationships with no lexical trigger

John killed the dog. It was threatening his chickens.

Connectives lexicalizing a means or result

John killed the dog.

Unspecified causal relationships

The treatment is linked to better outcomes.

slide-41
SLIDE 41

Actual corpus examples can get quite complex.

“For market discipline to effectively constrain risk, financial institutions must be allowed to fail.” Average causal sentence length: 30 words “If properly done, a market sensitive regulatory authority not only prevents some of the problems, but is pro-market, because we have investors now who are unwilling to invest even in things they should.”

slide-42
SLIDE 42

Complex corpus examples

slide-43
SLIDE 43

Outline

  • What is Surface Construction Labeling (SCL)?
  • The causality task (Dunietz 2018)
  • The Because onomasiological annotation scheme
  • Making the task doable

– Constructicon as an annotation tool – Overlapping Relations

  • Slow research musings about grammaticalization

and the contents of the constructicon

slide-44
SLIDE 44

Decision Tree

slide-45
SLIDE 45

Connective pattern <cause> prevents <effect> from <effect> <enough cause> for <effect> to <effect> Annotatable words prevent, from enough, for, to WordNet verb senses prevent.verb.01 prevent.verb.02 Type Verbal Complex Degree INHIBIT FACILITATE Type restrictions Not PURPOSE Example His actions prevented disaster.

There’s enough time for you to find a restroom.

Annotators were guided by a “constructicon.”

slide-46
SLIDE 46

Inter-annotator agreement

Causal Overlapping Connective spans (F1) 0.77 0.89 Relation types (κ) 0.70 0.91 Degrees (κ) 0.92 (n/a) CAUSE/ARGC spans (%) 0.89 0.96 CAUSE/ARGC spans (Jaccard) 0.92 0.97 CAUSE/ARGC heads (%) 0.92 0.96 EFFECT/ARGE spans (%) 0.86 0.84 EFFECT/ARGE spans (Jaccard) 0.93 0.92 EFFECT/ARGE heads (%) 0.95 0.89

260 sentences; 98 causal instances; 82 overlapping relations

slide-47
SLIDE 47

The Causal Language Constructicon

  • 290 construction variants
  • 192 lexically distinct connectives

– “prevent” and “prevent from” use the same primary connective word

slide-48
SLIDE 48
slide-49
SLIDE 49

New examples can be added

  • Annotators recommend new connectives
  • Do a quick corpus study to verify that that

connective frequently expresses causality

– In other words, its use for causality seems to be conventionalized

slide-50
SLIDE 50
slide-51
SLIDE 51

We annotate 7 different types

  • f overlapping relations.

TEMPORAL CORRELATION HYPOTHETICAL OBLIGATION/PERMISSION CREATION/TERMINATION EXTREMITY/SUFFICIENCY CONTEXT

After; once; during As; the more…the more… If…then… Require; permit Generate; eliminate So…that…; sufficient…to… Without; when (circumstances where…)

slide-52
SLIDE 52

Lingering difficulties with

  • ther overlapping relations

Origin/destination:

toward that goal

Topic:

fuming over recent media reports

Component:

as part of the liquidation

Evidentiary basis:

went to war on bad intelligence

Having a role:

as an American citizen

Placing in a position:

puts us at risk

slide-53
SLIDE 53

Overlapping Relation Examples

  • Temporal

– Within minutes after the committee released its letter, Torricelli took the senate floor to apologize to the people of New Jersey.

  • Correlation

– Auburn football players are reminded of last year’s losses every time they go into the weight room.

  • Hypothetical

– Previously, he allowed increases in emissions as long as they did not exceed the rate of economic growth.

slide-54
SLIDE 54

Overlapping Relation Examples

  • Obligation/Permission

– He will roll back a provision known as a new source review that compels utilities to install modern pollution controls whenever the significantly upgrade

  • lder plants.
  • “whenever” is also a connective
  • Creation/Termination

– Many expected synergies of financial service activities gave rise to conflicts and excessive risk taking.

  • Context

– With Hamas controlling Gaza, it was not clear that Mr. Abbas had the power to carry out his decrees.

slide-55
SLIDE 55

Annotators applied several tests to determine when an

  • verlapping relation was also causal.
  • Can the reader answer a “why” question?
  • Does the cause precede the effect?
  • Counterfactuality(Grivaz, 2010) : would the effect

have been just as probable without the cause?

Rules out: My bus will leave soon, I just finished my breakfast.

  • Ontological asymmetry (Grivaz, 2010):

could the cause and effect be reversed?

– Rules out: It is a triangle. It has three sides.

  • Can it be rephrased as “because?”
slide-56
SLIDE 56

Temporal with and without causal interpretation After last year’s fiasco, everyone is being cautious.

ARGE ARGC MOTIVATION + TEMPORAL

After last year’s fiasco, they’ve rebounded this year.

ARGE ARGC TEMPORAL

slide-57
SLIDE 57

Conditional hypotheticals don’t have to be causal, but most are.

84% carry causal meaning Non-causal: If he comes, he’ll bring his wife. Causal: If I told you, I’d have to kill you.

slide-58
SLIDE 58

Overlapping relations

slide-59
SLIDE 59

Causality has seeped into the temporal and hypothetical domains.

~7% are expressed as hypotheticals Of the causal expressions in the corpus: > 14% are piggybacked on temporal relations

slide-60
SLIDE 60

Conditional hypotheticals don’t have to be causal, but most are.

84% carry causal meaning Non-causal: If he comes, he’ll bring his wife. Causal: If I told you, I’d have to kill you.

slide-61
SLIDE 61

Outline

  • What is Surface Construction Labeling (SCL)?
  • The causality task (Dunietz 2018)
  • The Because onomasiological annotation scheme
  • Making the task doable

– Constructicon as an annotation tool – Overlapping Relations

  • Slow research musings about grammaticalization

and the contents of the constructicon

slide-62
SLIDE 62

Slow Research Musings

  • Usefulness of SCL and onomasiological

annotation

– In NLP tasks – For linguistic discovery

slide-63
SLIDE 63

Slow research musings

  • Convention vs exploitation

– Patrick Hanks, Lexical Analysis: Norms and Exploitations, MIT Press.

  • With respect to constructions is conventional the same

as grammaticalized?

  • In 84% of if-then in the Because corpus, the sentence

seems to have the intent of expressing causality:

– Does that mean that if-then is a conventional/grammaticalized way of expression causality? – If so, does it lead us to extreme conclusions…….

slide-64
SLIDE 64

Slow research musings

  • Constructions whose meaning side is a frame:

e.g., causality

  • Constructions whose meaning side is a

function: e.g., “to V-base”, “Det N”

slide-65
SLIDE 65

Slow Research Musings

  • Fusion constructions vs overlapping relations

– English Dative shift The child was hungry. So his mother gave him (recipient) a cookie (theme). So his mother gave a cookie (theme) to him (recipient).

  • Animacy, information status, definiteness, NP weight, change of

possession/information/state

  • Joan Bresnan, Anna Cueni, Tatiana Nikitina, and Harald Baayen. 2007. "Predicting the Dative

Alternation." In Cognitive Foundations of Interpretation, ed. by G. Boume, I. Kraemer, and J.

  • Zwarts. Amsterdam: Royal Netherlands Academy of Science, pp. 69--94.
  • Frishkoff et al. (2008) Principal components analysis reduced Bresnan’s 14 features to four,

roughly corresponding to verb class, NP weight, animacy, and information status.

slide-66
SLIDE 66

Slow research musings

Fusion Constructions

– Chinese ba (aside from issues of the part of speech)

  • proposition-level completedness, lexical aspect,

discourse information status of the direct object, affectedness of the direct object

slide-67
SLIDE 67

Slow research musings

  • What is the meaning side of fusional

constructions in a constructicon?

– An abstraction from which all the other functions follow?

  • English Dative Shift: Topic worthiness of the recipient
  • Ba: affectedness or completedness

– Or would fusional constructions be represented as a collection of overlapping relations?

slide-68
SLIDE 68

Stop here

slide-69
SLIDE 69

Lingering difficulties include

  • ther overlapping relations

and bidirectional relationships.

For us to succeed, we all have to cooperate.

succeed cooperate

enables

succeed cooperate

necessitates

slide-70
SLIDE 70
  • But not every construction evokes a frame.

There could be a function instead of a frame, like information structure (Goldberg). (Fillmore, Lee-Goldman, and Rhodes, 2012). Chuck mentioned the need for interactional frames for pragmatics and some people have pursued it.

slide-71
SLIDE 71

Foundation: Form and Meaning

  • Constructions (pairs of form and meaning) (Fillmore):

– Meaning is not a discrete space

  • However there are centroids of meaning that tend to get

grammaticalized

– Meaning is grammaticalized in arbitrary constellations of linguistic units – Meaning can be compositional or conventional (idiomatic)

  • Languages have comparable categories

– Common centroids of form and meaning like “noun”

  • Languages differ

– How they carve up the meaning space when they grammaticalize it – How they spread from the centroids – What constellations of linguistic units are used

slide-72
SLIDE 72

Meaning is not a discrete space

  • http://www.ecenglish.com/learnenglish/lessons/will-would-shall-should
  • Desire, preference, choice or consent:

– Will you please be quiet? – He won’t wash the dishes

  • Future:

– It will be a great party. – I will go to the market.

  • Capability/capacity:

– The ship will take three hundred guests. – This bottle will hold two litres of wine.

  • Other:

– Phone rings: That will be my son

slide-73
SLIDE 73

Past Present Future Hypo- thetical Counter- factual Prohibitive Imperative

Ferdinand De Haan, “On Representing Semantic Maps” The ovals represent the points in semantic space. The outlines each represent the irrealis morpheme in one language, showing what part of the semantic space it covers. Discrete centroids tend to be grammaticalized, but not in the same way in each language.

Muyuw irrealis morpheme Sinaugoro irrealis morpheme Manam irrealis morpheme Susurunga irrealis morpheme

slide-74
SLIDE 74

Related spaces are grammaticalized differently Anaphoricity, Specificity, Familiarity

Discourse-old

I met a student. The student was tall.

Discourse Predictable

I went to a wedding. The bride and groom looked great.

Specific

I’m looking for a

  • student. Her name is

Chris.

Non Specific

I’m looking for a

  • student. I need one

to help me with something.

Familiarity/Context

Hand me the pen on the desk.

Genericity

Uniqueness Abstract nouns h Mass nouns “The” and “a” are not

  • meanings. They are

grammaticalizations of parts of this space.

slide-75
SLIDE 75

Constructions: pairs of form and meaning (Goldberg)

  • Morpheme

– -ed past time

  • Word

– Student

  • Phrase (productive and meaning is compositional)

– The student (an anchored instance of a student)

slide-76
SLIDE 76

Constructions: pairs of form and meaning

  • Phrase (productive but meaning is non-compositional)

– What a nice shirt! – Why not read a book? (suggestion, invitation, deontic)

  • Phrase (not productive, Goldberg)

– To prison – To bed – To school – To hospital – On holiday – *to airport

– *To bath

  • Idiom

– Out in left field

slide-77
SLIDE 77

Constructions: pairs of form and meaning

  • Arbitrary constellation of units of form, possibly discontinuous,

possibly non-compositional

– What is she doing going to the movies? (Fillmore et al.) (incongruity) – What do you mean you don’t know? (disbelief, incongruity) – It was too big to fail – No intellectual characteristic is too ineffable for assessment.

slide-78
SLIDE 78

Task definition: connective discovery + argument identification

I worry because I care. I worry because I care.

Connective discovery

Find lexical triggers

  • f causal relations

Argument identification

Identify cause & effect spans for each connective

slide-79
SLIDE 79
  • Improve shallow semantic parsing coverage using richer, more

flexible linguistic representations, leading to a unified approach to Frame Semantic Parsing, integrating the FrameNet Lexicon and the FrameNet Constructicon.

  • Create a proof-of-concept for SCL:
  • Design annotation guidelines & annotate a corpus

using these representations.

  • Build automated machine learning taggers

for constructional realizations of semantic relations.

Goal of Surface Construction Labeling

Goal of the causal labeling project

slide-80
SLIDE 80

Four Automatic Labelers

  • Syntax-based

– Benchmark – CausewayS

  • String-based

– CausewayL

  • Neural: DeepCx
slide-81
SLIDE 81

Syntax-based connective discovery: each construction is treated as a partially-fixed parse tree fragment

worry/VBP

nsubj advcl

I/PRP care/VBP

mark nsubj

I/PRP because/IN I worry because I care.

“head” of because I care

slide-82
SLIDE 82

Syntax-based connective discovery: each construction is treated as a partially-fixed parse tree fragment.

advcl mark

because/IN

nsubj

I/PRP

nsubj

I/PRP worry/VBP care/VBP I worry because I care.

slide-83
SLIDE 83

Syntax-based connective discovery: each construction is treated as a partially-fixed parse tree fragment.

advcl mark

because/IN

slide-84
SLIDE 84

Benchmark: Syntax-based dependency path memorization heuristic.

Connective Parse paths to possible cause/effect heads Causal / Not causal prevent from nsubj, advcl 27/ 4 prevent from nsubj, advmod 0 / 8 because of case, case  nmod

14 / 1

… … …

slide-85
SLIDE 85

The syntax-based benchmark system

  • Benchmark

– For each causal tree fragment

  • Count how many times it is causal in the training data
  • Count how many times the same fragment is not causal in the

training data

  • If that pattern is causal more often in the training data, then

whenever you see it in the test data, label it as causal.

  • If that pattern is non-causal more often in the training data, then

whenever you see it in the test data, don’t label it as causal.

– High precision: If the benchmark system thinks it’s causal, it is usually right – Low recall: The benchmark system misses too many instances of causality

slide-86
SLIDE 86

Syntax-based connective discovery: TRegex patterns are extracted in training, and matched at test time.

Training: Test: I worry because I care.

advcl mark

because/IN

(/^because_[0-9]+$/ <2 /^IN.*/ <1 mark > (/.*_[0-9]+/ <1 advcl > (/.*_[0-9]+/))) (/^because_[0-9]+$/ <2 /^IN.*/ <1 mark > (/.*_[0-9]+/ <1 advcl > (/.*_[0-9]+/)))

I worry because I love you.

TRegex 1

I worry because I love you.

+

1 Levy and Andrew, 2006

slide-87
SLIDE 87

The syntax-based CausewayS system

  • Fancier way of extracting tree fragments from

the training data

– Dreyfus-Wagner, minimum-weight sub-tree

  • For each tree fragment extracted from the

training data, find all instances of it in the test data.

  • CausewayS, at this point, has high recall (finds

a lot of instances of causality) but low precision (a lot of what it finds isn’t right).

  • Apply a classifier to increase precision.
slide-88
SLIDE 88

Syntax-based argument ID: Argument heads are expanded to include most dependents.

nsubj advcl

I/PRP

mark nsubj

I/PRP because/IN care/VBP worry/VBP

slide-89
SLIDE 89

Syntax-based argument ID: Argument heads are expanded to include most dependents.

nsubj advcl

I/PRP

mark nsubj

because/IN care/VBP worry/VBP I/PRP

slide-90
SLIDE 90

CausewayL: constructions are matched by regular expressions over word lemmas.

Training: Test: I worry because I care. I worry because I love you.

regex

I worry because I love you.

+

(ˆ | )([ \ S]+ )+?(because/IN) ([ \ S]+ )+? (ˆ | )([ \ S]+ )+?(because/IN) ([ \ S]+ )+?

slide-91
SLIDE 91

CausewayL: Arguments are labeled by a conditional random field.

labels featurized words … … CAUSE EFFECT EFFECT

Features include information about:

  • Word
  • Connective
  • Relationship between word & connective

𝑞 𝐳 𝐲 = 1 𝑎 𝐲 ෑ

𝑢=1 𝑈

exp ෍

𝑙=1 𝐿

𝜄𝑙𝑔

𝑙 𝑧𝑢, 𝑧𝑢−1, 𝐲𝑢

slide-92
SLIDE 92

CausewayL

  • Recall is high (finds a lot of matches)
  • Precision is low (a lot of what it finds is wrong)
  • Use a classifier to raise precision
slide-93
SLIDE 93

DeepCx: Neural Net, LSTM

slide-94
SLIDE 94

Analysis of results

  • DeepCx is better than Causeway in these circumstances

– Ambiguity (entropy) of the causal connective.

  • Eg., to has entropy of 1

– F1 for DeepCx: 40.6% – F1 for Causeway: 25.5%

– Connectives whose part of speech is Adverb

  • Causeway missed all instances of “why” and “when”

– Design problem?

– Connectives whose part of speech is Noun

  • Fewer instances to train on

– More words in the causal connective

  • But CausewayL does pretty well too because it is matching a pattern over a

whole sentence whereas DeepCx is proceeding one word at a time. CausewayS is missing complex parse paths.

slide-95
SLIDE 95

Analysis of results

  • CausewayL and DeepCx do better than CausewayS on most
  • verlapping relations
  • Amount of training data

– DeepCx performs better at all amounts of training data – It appears to be better at generalizing across patterns (vs generalizing within patterns) – The gap between CausewayS and DeepCx remains constant across all amounts of training data – CausewayL increases fastest with more training data

  • Length of Cause and Effect arguments

– DeepCx and CausewayL are better than CausewayS as argument length increases

  • CausewayL uses CRF to label argument spans
  • CausewayS uses the dependency trees
slide-96
SLIDE 96

Next steps

  • Reproduce with other meanings with large

constructicons such as comparatives

  • Reproduce with multiple languages
  • Apply where the constructicon is small and

quirky too such as incongruity

slide-97
SLIDE 97

Related annotation schemes and labelers for causality

slide-98
SLIDE 98

Previous projects have struggled to annotate real-world causality.

“A person infected with a <e1>flu</e1>

<e2>virus</e2> strain develops antibodies against

it.”

Cause-Effect(e2, e1) = "true"

SemEval 2007 Task 4

(Girju et al., 2007)

Richer Event Descriptions

(O’Gorman et al., 2016; Croft et al., 2016)

We’ve allocated a budget to equip the barrier with electronic detention equipment.

BEFORE-PRECONDITIONS

CaTeRS

(Mostafazadeh et al., 2016)

slide-99
SLIDE 99

Existing shallow semantic parsing schemes include some elements of causal language.

Penn Discourse Treebank

(Prasad et al., 2008)

PropBank

(Palmer et al., 2005)

FrameNet

(Fillmore & Baker, 2010; Ruppenhofer et al., 2016)

CAUSATION

EFFECT EFFECT CAUSER

made bow me He

PURPOSE

to show his dominance .

slide-100
SLIDE 100

Others have focused specifically on causality.

Causality in TempEval-3

(Mirza et al., 2014)

CAUSE EVENT BEFORE EVENT TLINK

HP acquired 730,070 common shares as a result of a stock purchase agreement.

BioCause

(Mihaila et al., 2013)

CaTeRS

(Mostafazadeh et al., 2016)

Richer Event Description

(O’Gorman et al., 2016)

We’ve allocated a budget to equip the barrier with electronic detention equipment.

BEFORE-PRECONDITIONS

slide-101
SLIDE 101

A note on universals

  • Categories in linguistics:

– Degrees of universality

  • Describing a particular language on its own terms
  • Using comparable categories for language typology
  • Making a theory of universal categories for all

languages

slide-102
SLIDE 102

A note on universals

  • Universals in NLP

– Universal categories for cross-lingual training – Some new multi-lingual resources are going viral

  • Google universal parts of speech
  • Google Universal Dependencies (60 languages)
  • Abstract Meaning Representation
  • UniMorph

– They know not what they do. We need to be involved.

slide-103
SLIDE 103

Acknowledgements

  • Jesse Dunietz
  • Jaime Carbonell
  • David Mortensen
  • Archna Bhatia
  • Nathan Schneider
  • Miriam Petruck
  • Judith Klavans
slide-104
SLIDE 104

Thank you