lexical functional grammar
play

Lexical-Functional Grammar Ash Asudeh Carleton University - PowerPoint PPT Presentation

Lexical-Functional Grammar Ash Asudeh Carleton University University of Iceland July 3, 2009 1 Architecture and Structures 2 Basic Syntactic Architecture of LFG Two basic, simultaneous representations of syntax:


  1. Functional Descriptions and Subsumption • F-descriptions are true of not just the smallest, ‘intuitively intended’ f-structure, but also any larger f-structure that contains the same information.* * This relationship is called subsumption : In general, a structure A subsumes a structure B if and only if A and B are identical or B contains A and additional information not included in A. ‘ GO SUBJ ’ PRED TENSE FUTURE f subsumes g ‘ GO SUBJ ’ PRED ‘ PRO ’ PRED NUM SG SUBJ CASE NOM SUBJ NUM SG • An f-description is therefore true of not just the minimal f-structure that satisfies the description: the f-description is also true of the infinitely many other f-structures that the intended, minimal f-structure subsumes. 27

  2. Minimization • There is a general requirement on LFG’s solution algorithm that it yield the minimal solution: no features that are not mentioned in the f-description may be included. • Let’s look at an example from Dalrymple (2001). (1)David sneezed. Minimal consistent f-structure • F-description: subsumes (20) ( f PRED) ¼ ‘ SNEEZE h SUBJ i ’ ( f TENSE) ¼ PAST ( f SUBJ) ¼ g ( g PRED) ¼ ‘ DAVID ’ C o n s i s t e n t b u t n o n - m i n i m a l f - s t r u c t u r e 28

  3. Lexical Generalizations in LFG V (2) yawns ( PRED )=‘yawn SUBJ ’ ( VFORM )= FINITE ( TENSE )= PRES ( SUBJ PERS )=3 ( SUBJ NUM )= SG A lot of this f-description is shared by other verbs. 29

  4. LFG Templates: Relations between Descriptions (2) yawns ( PRED )=‘yawn SUBJ ’ (3) PRESENT = ( VFORM )= FINITE ( VFORM )= FINITE ( TENSE )= PRES ( TENSE )= PRES ( SUBJ PERS )=3 3 SG = ( SUBJ PERS )=3 ( SUBJ NUM )= SG ( SUBJ NUM )= SG ⇧ (4) yawns ( PRED )=‘yawn SUBJ ’ @ PRESENT @3 SG 30

  5. Templates: Factorization and Hierarchies (5) FINITE = ( VFORM )= FINITE (7) 3P ERSON S UBJ = ( SUBJ PERS )=3 PRES - TENSE = ( TENSE )= PRES S ING S UBJ = ( SUBJ NUM )= SG 3S G = @3P ERSON S UBJ = @ FINITE PRESENT @S ING S UBJ @ PRES - TENSE ⇧ ⇧ 3P ERSON S UBJ S ING S UBJ PRES - TENSE FINITE 3 SG PRESENT 31

  6. Templates: Factorization and Hierarchies (9) P RES 3S G = @ PRESENT (4) yawns ( PRED )=‘yawn SUBJ ’ @3 SG @ PRESENT @3 SG ⇧ 1) yawns ( PRED )=‘yawn SUBJ ’ @P RES 3S G PRES - TENSE 3P ERSON S UBJ S ING S UBJ FINITE 3 SG PRESENT P RES 3S G 32

  7. Templates: Boolean Operators (15) P RES N OT 3 SG = @ PRESENT (16) ( VFORM )= FINITE @3 SG ⇧ ( TENSE )= PRES ( SUBJ PERS )=3 Negation ( SUBJ NUM )= SG PRES - TENSE 3P ERSON S UBJ S ING S UBJ FINITE 3 SG PRESENT P RES N OT 3 SG P RES 3S G 33

  8. Hierarchies: Templates vs. Types • Type hierarchies are and/or lattices: (1) HEAD • Motherhood: or NOUN RELATIONAL • Multiple Dominance: and C - NOUN GERUND VERB • Type hierarchies encode inclusion/inheritance and place constraints on how the inheritance is interpreted. • LFG template hierarchies encode only inclusion: multiple dominance not interpreted as conjunction, no real status for motherhood. • LFG hierarchies relate descriptions only: mode of combination (logical operators) is determined contextually at invocation or is built into the template. • HPSG hierarchies relate first-class ontological objects of the theory. • LFG hierarchies are abbreviatory only and have no real ontological status. 34

  9. Hierarchies: Templates vs. Types (1) HEAD HPSG NOUN RELATIONAL C - NOUN GERUND VERB PRES - TENSE 3P ERSON S UBJ S ING S UBJ FINITE LFG 3 SG PRESENT P RES N OT 3 SG P RES 3S G 35

  10. Parameterized Templates (12) INTRANSITIVE ( P ) = ( PRED )=‘ P SUBJ ’ 1) yawns ( PRED )=‘yawn SUBJ ’ @P RES 3S G ⇧ (13) yawns @ INTRANSITIVE (yawn) @P RES 3S G 36

  11. Parameterized Templates (18) TRANSITIVE ( P ) = ( PRED )=‘ P SUBJ , OBJ ’ (19) TRANS - OR - INTRANS ( P ) = @ TRANSITIVE ( P ) @ INTRANSITIVE ( P ) (20) ( PRED )=‘ eat SUBJ , OBJ ’ ( PRED )=‘ eat SUBJ ’ 37

  12. Temple Hierarchy with Lexical Leaves 3P ERSON S UBJ S ING S UBJ 3 SG PRESENT INTRANSITIVE TRANSITIVE P RES 3 SG TRANS - OR - INTRANS falls bakes cooked yawns eats 38

  13. Defaults in LFG The f-structure must have case and (23) ( CASE ) ( CASE )= NOM if nothing else provides its case, then its case is nominative. Paramerized template for defaults. (24) DEFAULT ( D V ) = D = V D Also illustrates that parameterized templates can have multiple arguments ⇧ (25) @ DEFAULT (( CASE ) NOM ) 39

  14. C-structure Annotation of Templates a. ADJUNCT ( P ) = ( ADJUNCT ) VP V ADVP* @ ADJUNCT - TYPE ( P ) = ( ADJUNCT ) ( ADJUNCT - TYPE )= VP - ADJ b. ADJUNCT - TYPE ( P ) = ( ADJUNCT - TYPE )= P ⇧ VP V ADVP* = @ ADJUNCT ( VP - ADJ ) 40

  15. Features in the Minimalist Program 41

  16. Features and Explanation • The sorts of features that are associated with functional heads in the Minimalist Program are well-motivated morphosyntactically, although other theories may not draw the conclusion that this merits phrase structural representation (cf. Blevins 2008). • Care must be taken to avoid circular reasoning in feature theory: • The ‘strong’ meta-feature: “This thing has whatever property makes things displace, as evidenced by its displacement.” • The ‘weak’ meta-feature: “This thing lacks whatever property makes things displace, as evidenced by its lack of displacement.” • The EPP feature: “This thing has whatever property makes things move to subject position, as evidenced by its occupying subject position.” 42

  17. Features and Simplicity • Adger (2003, 2008) considers three kinds of basic features: • Privative, e.g. [singular] • Binary, e.g. [singular +] • Valued, e.g. [number singular] • Adger considers the privative kind the simplest in its own right. • This may be true, but only if it does not introduce complexity elsewhere in the system (Culicover & Jackendoff 2005: ‘honest accounting’). • Notice that only the final type of feature treats number features as any kind of natural class within the theory (as opposed to meta- theoretically). 43

  18. Kinds of Feature-Value Combinations • Adger (2003): • Privative • [singular], [V], ... • Binary • [singular: +] (?) • Attribute-value • [Tense: past] 44

  19. Interpreted vs. Uninterpreted Features • Interpreted features: • [F] • Uninterpreted features: • [ u F] • All uninterpreted features must be eliminated (‘checked’). • Interpreted features are interpreted by the semantics. • Presupposes an interpretive (non-combinatorial) semantics. [ Notation from Adger 2003] 45

  20. Feature Strength • Strong features must be checked locally: Trigger Move/Internal Merge/Remerge • [F*] • Weak features do not have to be checked locally: Do not trigger Move • [F] [Notation from Adger 2003] 46

  21. An Example: Auxiliaries • Adger (2003:181) “When [ u Infl: ] on Aux is valued by T, the value is strong; when [ u Infl: ] on v is valued by T, the value is weak.” TP Subject T T[past] NegP Neg v P � Subject � v Verb + v [ u Infl]... 47

  22. Locality of Feature Matching • Adger (2003:218) Locality of Matching Agree holds between a feature F on X and a matching feature F on Y if and only if there is no intervening Z[F]. Intervention In a structure [X ... Z ... Y], Z intervenes between X and Y iff X c- commands Z and Z c-commands Y. 48

  23. Feature-Value Unrestrictiveness & Free Valuation • Asudeh & Toivonen (2006) argue that the Minimalist feature system of Adger (2003) has two undesirable properties. Feature-value unrestrictiveness Feature valuation is unrestricted with respect to what values a valued feature may receive. Free valuation Feature valuation appears freely, subject to locality conditions. • This results in a very unconstrained theory of features. • This may sound good, because it’s less stipulative and hence more Minimal, but from a theory perspective it is bad: unconstrained theories are less predictive. 49

  24. Example: English Subject Agreement (1) Gilgamesh missed Enkidu (2) Gilgamesh misses Enkidu TP TP T[singular] T[past] v P v P Gilgamesh Gilgamesh v v v VP v VP miss v [ u Infl:singular] � miss � NP miss v [ u Infl:past] � miss � NP Enkidu Enkidu • Contrast with HPSG: MP has no typing of values (feature value unrestrictiveness) • Contrast with LFG: MP has valuation without specification (free valuation) 50

  25. Two Contrasting Feature Theories • HPSG (Pollard & Sag 1994): features are not just valued, the values are also typed • If two values can unify, they must be in a typing relation (one must be a subtype of the other). • Feature values in HPSG are thus tightly restricted by types. • LFG (Kaplan & Bresnan 1982, Bresnan 2001): features are not restricted, but there is no free valuation • A feature cannot end up with a given value unless there is an explicit equation in the system. 51

  26. Feature Simplicity and Constraint Types • LFG offers the opportunity to consider Adger’s three feature types in light of a single feature type, with varying constraint types. • LFG features are valued ( f is an LFG f(unctional)-structure): � � singular f NUMBER • Types of LFG feature constraints. • Defining equation: ( f NUMBER ) = singular • Existential constraint: ( f NUMBER ) • Negative existential constraint: ¬ ( f NUMBER ) • Constraining equation: ( f NUMBER ) = c singular • Negative constraining equation: ( f NUMBER ) � = singular 52

  27. Feature Simplicity and Constraint Types • All features treated as valued features: no restriction on constraint types • All features treated as binary features: only positive and negative constraining equations allowed • All features treated as privative: only negative and existential constraints allowed • This understanding of privative features actually does treat number as a natural class. • This treats the notion of feature simplicity as a kind of meta- theoretical statement in an explicit, non-ad-hoc feature theory. 53

  28. Control and Raising 54

  29. Lexical Entries tried V ( ↑ PRED ) = ‘try � SUBJ , XCOMP � ’ ( ↑ SUBJ ) = ( ↑ XCOMP SUBJ ) seemed V ( ↑ PRED ) = ‘seem � CF � SUBJ ’ { ( ↑ SUBJ ) = ( ↑ XCOMP SUBJ ) | ( ↑ SUBJ PRONTYPE ) = EXPLETIVE ( ↑ SUBJ FORM ) = IT ( ↑ COMP ) } 55

  30. Raising to Subject/Subject Control C-structure IP ( ↑ SUBJ ) = ↓ ↑ = ↓ I ′ NP ↑ = ↓ Gonzo VP ↑ = ↓ ↑ = ↓ V 0 VP seemed/tried ↑ = ↓ ↑ = ↓ V 0 VP to ↑ = ↓ V 0 leave 56

  31. F-structures 57

  32. Copy Raising 58

  33. Data (1)Thora seems like she enjoys hot chocolate. (2)Thora seems like Isak pinched her again. (3)Thora seems like Isak ruined her book. (4)* Thora seems like Isak enjoys hot chocolate. (5)* Thora seems like Isak pinched Justin again. (6)* Thora seems like Isak ruined Justin’s book. 59

  34. Data (7)It seems like there is a problem here. (8)It seems like Thora is upset. (9)It seems like it rained last night. (10) There seems like there’s a problem here. (11) * There seems like it rained last night. 60

  35. Lexical Entries P 0 like 1 ( ↑ PRED ) = ‘like � SUBJ , COMP � ’ P 0 like 2 ( ↑ PRED ) = ‘like � CF � SUBJ ’ { ( ↑ SUBJ ) = ( ↑ XCOMP SUBJ ) | ( ↑ SUBJ PRONTYPE ) = EXPLETIVE ( ↑ SUBJ FORM ) = IT ( ↑ COMP ) } 61

  36. C-structure IP ( ↑ SUBJ ) = ↓ ↑ = ↓ DP I � ↑ = ↓ Richard VP ↑ = ↓ ( ↑ XCOMP ) = ↓ V 0 PP seems ↑ = ↓ seems / smells P � ( ↑ COMP ) = ↓ ↑ = ↓ P 0 IP ( ↑ SUBJ ) = ↓ ↑ = ↓ like DP I � ↑ = ↓ he VP smokes 62

  37. F-structure   ‘seem/smell’ ‘smoke’ ‘seem/ PRED     SUBJ         ‘like’ PRED       � �     ‘Richard’ SUBJ   PRED               ‘smoke’   PRED                 XCOMP   ‘pro’   PRED                   COMP 3     PERS           SUBJ           sg     NUM                     masc  GEND  63

  38. C-structure IP ( ↑ SUBJ ) = ↓ ↑ = ↓ DP I � ↑ = ↓ There VP ( ↑ XCOMP ) = ↓ ↑ = ↓ V 0 PP ( ↑ XCOMP ) = ↓ ↑ = ↓ seemed P 0 IP ( ↑ SUBJ ) = ↓ ↑ = ↓ like DP I � there was a problem 64

  39. F-structure   ‘seem’ PRED     SUBJ         ‘like’ PRED           SUBJ               ‘be’   PRED           XCOMP  � �      there SUBJ  EXPL              XCOMP           ‘problem’   PRED           OBJ       � �       ‘a’  SPEC  PRED   65

  40. Unbounded Dependencies 66

  41. Filler-Gap Dependencies 67

  42. Functional Uncertainty • The syntactic relationship between the top and bottom of an unbounded dependency is represented with a functional uncertainty: • Top = MiddlePath-Func-Uncertainty Bottom-Func-Uncertainty (1) [What] [did Kim claim that Sandy suspected that Robin knew] [ ] top middle bottom ( ↑ FOCUS ) = ( ↑ COMP ∗ { OBJ | OBJ θ } ) top middle bottom (2) [What] [did Kim claim that Sandy suspected that Robin gave Bo] [ ] 68

  43. Wh -Questions: Example CP ‘ PRO ’ PRED NP C FOCUS PRONTYPE WH N C IP Q Who does NP I ‘ LIKE SUBJ , OBJ ’ PRED ‘D AVID ’ N VP PRED SUBJ David V OBJ like 69

  44. Wh -Questions: Annotated PS Rule ) CP QuesP C ( FOCUS ) = = ( FOCUS ) = ( QF OCUS P ATH ) ( Q ) = ( FOCUS W H P ATH ) ( Q PRONTYPE ) WH 70

  45. Wh -Questions: QuesP Metacategory QuesP NP PP AdvP AP (1)NP: Who do you like? (2)PP: To whom did you give a book? (3)AdvP: When did you yawn? (4)AP: How tall is Chris? 71

  46. Wh -Questions: Unbounded Dependency Equation English QF OCUS P ATH : XCOMP ADJ GF GF COMP OBJ ( LDD ) ( TENSE ) ( TENSE ) 72

  47. Wh -Questions: Pied Piping ) English W H P ATH : SPEC OBJ (1)[Whose book] did you read? (2)[Whose brother’s book] did you read? (3)[In which room] do you teach? 73

  48. Relative Clauses: Example 26) a man who Chris saw ‘ MAN ’ PRED ‘ A ’ PRED SPEC ‘ PRO ’ PRED TOPIC PRONTYPE REL RELPRO ADJ ‘ SEE SUBJ , OBJ ’ PRED NP ‘C HRIS ’ SUBJ PRED Det N OBJ a N CP N NP C man N IP who NP I N VP Chris V saw 74

  49. Relative Clauses: Annotated PS Rule ) CP RelP C ( TOPIC ) = = ( TOPIC ) = ( RT OPIC P ATH ) ( RELPRO ) = ( TOPIC R EL P ATH ) ( RELPRO PRONTYPE ) REL 75

  50. Relative Clauses: RelP Metacategory RelP NP PP AP AdvP (1)NP: a man who I selected (2)PP: a man to whom I gave a book (3)AP: the kind of person proud of whom I could never be (4)AdvP: the city where I live 76

  51. Relative Clauses: Unbounded Dependency Equation English RT OPIC P ATH : XCOMP ADJ GF GF COMP OBJ ( LDD ) ( TENSE ) ( TENSE ) 77

  52. Relative Clauses: Pied Piping (1) the man [who] I met ) English R EL P ATH : (2) the man [whose book] I read (3) the man [whose brother’s book] I SPEC OBL OBJ read (4) the report [the cover of which] I designed (5) the man [faster than whom] I can run (6) the kind of person [proud of whom] I could never be (7) the report [the height of the lettering on the cover of which] the government prescribes 78

  53. Relative Clauses: Pied Piping Example (27) a man whose book Chris read ‘ MAN ’ PRED ‘ A ’ PRED SPEC ‘ PRO ’ PRED SPEC PRONTYPE REL TOPIC ‘ BOOK ’ PRED RELPRO ADJ ‘ READ SUBJ , OBJ ’ PRED ‘C HRIS ’ PRED SUBJ OBJ NP Det N a N CP N NP C man Det N IP whose N NP I book N VP Chris V read 79

  54. Constraints on Extraction 80

  55. Empty Category Principle/ That -Trace (1)Who do you think [__ left]? (2)* Who do you think [that __ left]? (3)* What do you wonder [if __ smells bad]? (4)Who do you think [__ should be trusted]? (5)* Who do you think [that __ should be trusted]? (6)Who do you think [that, under no circumstances, __ should be trusted]? (7)Who do you wonder [if, under certain circumstances, __ could be trusted]? 81

  56. That- Trace in LFG • LFG has a relation called f-precedence that uses the native precedence of c-structure to talk about precedence between bits of f-structure. • F-precedence relies on LFG’s projection architecture and the inverse of the c-structure–f-structure mapping function ϕ . • The inverse is written ϕ -1 and returns the set of c-structure nodes that map to its argument f-structure node. F-precedence An f-structure f f-precedes an f-structure g ( f < f g ) if and only if for all n 1 ∈ ϕ -1 ( f ) and for all n 2 ∈ ϕ -1 ( g ), n 1 c-precedes n 2 . 82

  57. That- Trace in LFG • We can leverage LFG’s projection architecture to capture the fact that That- Trace is a ‘surfacy’ phenomenon (cf. ECP as a PF constraint in recent Minimalism). Form φ ... ... π • • • string c-structure f-structure 83

  58. That- Trace in LFG • Assume a native precedence relation on strings, yielding a notion of element that is string-adjacent to the right (‘next string element’), which we define as Right string ( π -1 ( * )), where * designates the current c-structure node in a phrase structure rule element or lexical entry. • Let’s abbreviate the right string-adjacent element to * as ≻ . • The semantics of ≻ is ‘the string element that is right string- adjacent to me’. • Note that π -1 returns string elements, not sets of string elements, because π is bijective, since c-structures are trees. 84

  59. That- Trace in LFG • We can use f-precedence and ≻ to capture the surfacy nature of That- Trace. • Basically, English has a (somewhat arbitrary) constraint that the right-adjacent string element to the complementizer must be locally realized. • This can be stated by requiring that any unbounded dependency function in the f-structure corresponding to the element that occurs in the string immediately after the complementizer should not f-precede the complementizer’s f-structure. 85

  60. Left Branch Constraint (1)Whose car did you drive __? (2)* Whose did you drive [__ car]? 86

  61. Left Branch Constraint in LFG • Do not include SPEC/POSS in GFs of possible extraction sites. • Note that the equation we looked at previously already disallows the extraction from passing through a SPEC in the first part. • We modify the equation as follows English QF OCUS P ATH : GF − SPEC } XCOMP ADJ GF COMP OBJ ( LDD ) ( TENSE ) ( TENSE ) 87

  62. Wh -Islands in LFG: Off-Path Constraints • The off-path metavariable ← refers to the f-structure that contains the attribute that the constraint is attached to. • The off-path metavariable → refers to the f-structure that is the value of the attribute that the constraint is attached to. English QF OCUS P ATH : GF − SPEC } XCOMP ADJ GF COMP OBJ ( LDD ) ( TENSE ) ( TENSE ) • Use ← to state the bottom cannot be in an f-structure that has an unbounded dependency function UDF , where UDF = {TOPIC | FOCUS}. English QF OCUS P ATH : GF − SPEC } XCOMP ADJ GF COMP OBJ ( LDD ) ( TENSE ) ( TENSE ) ¬ ( ← UDF ) 88

  63. Successive Cyclic Effects 89

  64. Successive Cyclicity • Data from languages such as Irish and Chamorro, which show successive marking along the extraction path, have motivated the claim that extraction/movement is ‘cyclic’ (not all at once). Cf. Phases in Minimalism. • Of course, this data does not argue for movement per se, as some have wrongly assumed, but rather that unbounded dependencies should 1. Be made up of a series of local relations; or 2. Have a way to refer to their environments as the dependency is constructed. • HPSG has adopted the first approach, LFG the second. 90

  65. Data: Irish ) a. Shíl mé goN mbeadh sé ann thought I would-be he there PRT • Note: Date from McCloskey I thought that he would be there. via Bouma et al. (2001). b. Dúirt mé gurL shíl mé goN mbeadh sé ann said I goN + PAST thought I would-be he there PRT I said that I thought that he would be there. c. an fear aL shíl mé aL bheadh ann [the man] j thought I would-be there PRT PRT j the man that I thought would be there d. an fear aL dúirt mé aL shíl mé aL bheadh ann [the man] j said I thought I would-be there PRT PRT PRT j The man that I said I thought would be there e. an fear aL shíl goN mbeadh sé ann [the man] j thought would-be he there PRT PRT j the man that thought he would be there 91

  66. Irish Successive Cyclicity in LFG ˆ aL C ( ↑ UDF ) = ( ↑ GF ) CF ∗ ( → UDF ) = ( ↑ UDF ) Note: UDF = {TOPIC | FOCUS}, CF = {XCOMP | COMP} ˆ goN C ( ↑ TENSE ) ¬ ( ↑ UDF ) 92

  67. Glue Semantics 93

  68. Glue Semantics • Glue Semantics is a type-logical semantics that can be tied to any syntactic formalism that supports a notion of headedness. • Glue Semantics can be thought of as categorial semantics without categorial syntax. • The independent syntax assumed in Glue Semantics means that the logic of composition is commutative , unlike in Categorial Grammar. • Selected works: Dalrymple (1999, 2001), Crouch & van Genabith (2000), Asudeh (2004, 2005a,b, in prep.), Lev 2007, Kokkonidis (in press) 94

  69. Glue Semantics • Lexically-contributed meaning constructors := M : G Meaning language term Composition language term • Meaning language := some lambda calculus • Model-theoretic • Composition language := linear logic • Proof-theoretic • Curry Howard Isomorphism between formulas (meanings) and types (proof terms) • Successful Glue Semantics proof: Γ � M : G t 95

  70. Key Glue Proof Rules with Curry-Howard Terms Abstraction : Implication Introduction Application : Implication Elimination [ x : A ] 1 · · · · · · · a : A f : A � B · · f : B � E f ( a ) : B � I , 1 λ x . f : A � B Pairwise Conjunction Substitution : Elimination [ x : A ] 1 [ y : B ] 2 Beta reduction for let : · · let a × b be x × y in f f [ a / x , b / y ] ⇒ β · · · · a : A ⊗ B f : C ⊗ E , 1 , 2 let a be x × y in f : C 96

  71. Example: Mary laughed 1. mary : ↑ σ e   ‘laugh � SUBJ � ’ PRED f � � 2. laugh : ( ↑ SUBJ ) σ e ⊸ ↑ σ t   g ‘Mary’ SUBJ PRED 1 ′′ . mary : m 1 ′ . mary : g σ e 2 ′ . laugh : g σ e ⊸ f σ t 2 ′′ . laugh : m ⊸ l Proof Proof 1. mary : m Lex. Mary mary : m laugh : m ⊸ l ≡ 2. laugh : m ⊸ l Lex. laughed ⊸ E laugh ( mary ) : l 3. laugh ( mary ) : l E ⊸ , 1, 2 97

  72. Example: Most presidents speak 1. λ R λ S . most ( R , S ) : ( v ⊸ r ) ⊸ ∀ X . [( p ⊸ X ) ⊸ X ] Lex. most 2. president ∗ : v ⊸ r Lex. presidents 3. speak : p ⊸ s Lex. speak president ∗ : λ R λ S . most ( R , S ) : ( v ⊸ r ) ⊸ ∀ X . [( p ⊸ X ) ⊸ X ] v ⊸ r λ S . most ( president ∗ , S ) : speak : ∀ X . [( p ⊸ X ) ⊸ X ] p ⊸ s ⊸ E , [ s/X ] most ( president ∗ , speak ) : s 98

  73. Example: Most presidents speak at least one language   Single parse ‘speak � SUBJ , OBJ � ’ PRED   ‘president’   PRED   SUBJ   ➡ � �   ‘most’   SPEC PRED   Multiple scope possibilities       ‘language’ PRED (Underspecification through     OBJ � �     quantification) ‘at-least-one’ SPEC PRED 1. λ R λ S . most ( R , S ) : Lex. most ( v1 ⊸ r1 ) ⊸ ∀ X . [( p ⊸ X ) ⊸ X ] 2. president ∗ : v1 ⊸ r1 Lex. presidents 3. speak : p ⊸ l ⊸ s Lex. speak 4. λ P λ Q . at - least - one ( P , Q ) : Lex. at least one ( v2 ⊸ r2 ) ⊸ ∀ Y . [( l ⊸ Y ) ⊸ Y ] 5. language : v2 ⊸ r2 Lex. language 99

  74. Most presidents speak at least one language Subject wide scope λ P λ Q . a - l - o ( P , Q ) : lang : λ x λ y . speak ( x , y ) : [ z : p ] 1 ( v2 ⊸ r2 ) ⊸ ∀ Y . [( l ⊸ Y ) ⊸ Y ] v2 ⊸ r2 p ⊸ l ⊸ s λ Q . a - l - o ( lang , Q ) : λ y . speak ( z , y ) : president ∗ : λ R λ S . most ( R , S ) : ∀ Y . [( l ⊸ Y ) ⊸ Y ] l ⊸ s ( v1 ⊸ r1 ) ⊸ ∀ X . [( p ⊸ X ) ⊸ X ] v1 ⊸ r1 [ s/Y ] a - l - o ( lang , λ y . speak ( z , y )) : s λ S . most ( president ∗ , S ) : ⊸ I , 1 ∀ X . [( p ⊸ X ) ⊸ X ] λ z . a - l - o ( lang , λ y . speak ( z , y )) : p ⊸ s [ s/X ] most ( president ∗ , λ z . a - l - o ( lang , λ y . speak ( z , y ))) : s 100

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend