Natural Language Generation Demos Basics of NLG NLG concepts - - PowerPoint PPT Presentation

natural language generation
SMART_READER_LITE
LIVE PREVIEW

Natural Language Generation Demos Basics of NLG NLG concepts - - PowerPoint PPT Presentation

Natural Language Generation Scott Farrar CLMA, University of Washington far- rar@u.washington.edu Natural Language Generation Demos Basics of NLG NLG concepts Issues in NLG NLG subtasks Scott Farrar Architecture of NLG systems CLMA,


slide-1
SLIDE 1

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Natural Language Generation

Scott Farrar CLMA, University of Washington farrar@u.washington.edu March 3, 2010

1/50

slide-2
SLIDE 2

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Today’s lecture

1

Demos

2

Basics of NLG NLG concepts Issues in NLG NLG subtasks

3

Architecture of NLG systems Two-step architecture Three-step architecture

4

Hw7

2/50

slide-3
SLIDE 3

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Personals

Fun, loving woman looking to be a seat warmer on occasion this spring and summer. You be over 45, not married, and not a gang type biker. Want someone safe, sane and

  • responsible. I like Harley’s or a decent rice grinder. No

crotch rockets. Me: 5’6” HWP 120lb. love to ride, love the sun, love life, love to laugh. I really like men with beards/mustaches/goatees. :o)

3/50

slide-4
SLIDE 4

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Personals

Fun, loving woman looking to be a seat warmer on occasion this spring and summer. You be over 45, not married, and not a gang type biker. Want someone safe, sane and

  • responsible. I like Harley’s or a decent rice grinder. No

crotch rockets. Me: 5’6” HWP 120lb. love to ride, love the sun, love life, love to laugh. I really like men with beards/mustaches/goatees. :o) I am educated, stable, employed, 6ft tall, 200 lbs, brown eyes and brown hair. I work long hours at times, im sensitive at times, like good music, food, and conversations. I am a good friend, a good person, and easy to talk to. I’m looking for a slender asian girl who is educated and motivated/passionate. I like a girl who is educated. If you are out there, I would like to get to know you. Your pic gets mine.

3/50

slide-5
SLIDE 5

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG demo for pollen count1

Area Pollen Value North 1 North West 1 Central 1 North East 1 South West 1 South East 1 Grass pollen levels for Saturday remain at the low levels of recent weeks with values of 1 across the whole country. Therefore only those most sensitive to pollen will be affected.

1http://www.aclweb.org/aclwiki/index.php?title=Online NLG demos 4/50

slide-6
SLIDE 6

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG demo for pollen count

Area Pollen Value North 1 North West 1 Central 1 North East 1 South West 1 South East 10

5/50

slide-7
SLIDE 7

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG demo for pollen count

Area Pollen Value North 1 North West 1 Central 1 North East 1 South West 1 South East 10 Grass pollen levels for Saturday remain at the low levels of recent weeks with values of 1 across most parts of the

  • country. However, in South Eastern areas, pollen levels will

be very high with values of 10.

5/50

slide-8
SLIDE 8

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG demo for pollen count

Area Pollen Value North 1 North West 1 Central 1 North East 1 South West 1 South East 10 Grass pollen levels for Saturday remain at the low levels of recent weeks with values of 1 across most parts of the

  • country. However, in South Eastern areas, pollen levels will

be very high with values of 10.

5/50

slide-9
SLIDE 9

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG demo for pollen count

Area Pollen Value North 10 North West 10 Central 10 North East 10 South West 10 South East 10

6/50

slide-10
SLIDE 10

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG demo for pollen count

Area Pollen Value North 10 North West 10 Central 10 North East 10 South West 10 South East 10 Grass pollen levels for Saturday have increased from the low levels of recent weeks with values of 10 across the whole

  • country. Therefore making Saturday a particularly

unpleasant day for hay fever sufferers.

6/50

slide-11
SLIDE 11

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG demo for pollen count

Area Pollen Value North 10 North West 10 Central 10 North East 10 South West 10 South East 10 Grass pollen levels for Saturday have increased from the low levels of recent weeks with values of 10 across the whole

  • country. Therefore making Saturday a particularly

unpleasant day for hay fever sufferers.

6/50

slide-12
SLIDE 12

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Today’s lecture

1

Demos

2

Basics of NLG NLG concepts Issues in NLG NLG subtasks

3

Architecture of NLG systems Two-step architecture Three-step architecture

4

Hw7

7/50

slide-13
SLIDE 13

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

What is NLG?

Definition

Natural language generation is a CL subfield with the aim

  • f producing meaningful, grammatical utterances in natural

language from some non-linguistic input. The NLG process is based on some communicative goal (e.g., refute, describe, agree), and according to some larger discourse plan.

8/50

slide-14
SLIDE 14

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Specific Goals

Goal: describePerson SAM

(subclass Family Collection) (subclass Human LivingThing) (subclass MaleHuman Human) (instance Sam MaleHuman) (inFamily Sam f1) (familyName f1 Smith) (age Sam 32) (maritalStatus Sam single) (know Sam Jill) 9/50

slide-15
SLIDE 15

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Specific Goals

Goal: describePerson SAM

(subclass Family Collection) (subclass Human LivingThing) (subclass MaleHuman Human) (instance Sam MaleHuman) (inFamily Sam f1) (familyName f1 Smith) (age Sam 32) (maritalStatus Sam single) (know Sam Jill)

Sam Smith is 32 years old and is single.

9/50

slide-16
SLIDE 16

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Specific Goals

Goal: refute PROP45 given: like(BILL, SALLY )

10/50

slide-17
SLIDE 17

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Specific Goals

Goal: refute PROP45 given: like(BILL, SALLY ) It’s not true that Bill likes Sally. Bill doen’t like Sally. Bill likes Sally, right!

10/50

slide-18
SLIDE 18

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Specific Goals

Goal: (compare WA, VA) Given: loc(WA, NORTHWEST) ∧ loc(VA, EASTCOAST)

11/50

slide-19
SLIDE 19

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Specific Goals

Goal: (compare WA, VA) Given: loc(WA, NORTHWEST) ∧ loc(VA, EASTCOAST) Washington State is located in the northwest, while Virginia is on the east coast.

11/50

slide-20
SLIDE 20

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Specific Goals

Goal: (compare WA, VA) Given: loc(WA, NORTHWEST) ∧ loc(VA, EASTCOAST) Washington State is located in the northwest, while Virginia is on the east coast. Washington State is located in the northwest and Virginia on the east coast.

11/50

slide-21
SLIDE 21

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Some generation applications

12/50

slide-22
SLIDE 22

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Some generation applications

The output component of a machine translation system

12/50

slide-23
SLIDE 23

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Some generation applications

The output component of a machine translation system Interface to a database system

12/50

slide-24
SLIDE 24

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Some generation applications

The output component of a machine translation system Interface to a database system Interface to an expert system (math tutor)

12/50

slide-25
SLIDE 25

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Some generation applications

The output component of a machine translation system Interface to a database system Interface to an expert system (math tutor) Autogeneration of help pages for a software system

12/50

slide-26
SLIDE 26

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Some generation applications

The output component of a machine translation system Interface to a database system Interface to an expert system (math tutor) Autogeneration of help pages for a software system Robot-human communication

12/50

slide-27
SLIDE 27

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Some generation applications

The output component of a machine translation system Interface to a database system Interface to an expert system (math tutor) Autogeneration of help pages for a software system Robot-human communication Data summarization (stock/weather reports)

12/50

slide-28
SLIDE 28

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Some generation applications

The output component of a machine translation system Interface to a database system Interface to an expert system (math tutor) Autogeneration of help pages for a software system Robot-human communication Data summarization (stock/weather reports) Text summarization

12/50

slide-29
SLIDE 29

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Comparison to NLU

Similarities

13/50

slide-30
SLIDE 30

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Comparison to NLU

Similarities

both utilize a lexicon, grammar, and knowledge representation

13/50

slide-31
SLIDE 31

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Comparison to NLU

Similarities

both utilize a lexicon, grammar, and knowledge representation have same “endpoints” (internal computational representation, natural language utterances).

13/50

slide-32
SLIDE 32

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Comparison to NLU

Similarities

both utilize a lexicon, grammar, and knowledge representation have same “endpoints” (internal computational representation, natural language utterances). both symbolic and stochastic (corpus-based) methods apply

13/50

slide-33
SLIDE 33

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Comparison to NLU

Differences

14/50

slide-34
SLIDE 34

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Comparison to NLU

Differences

NLU is about hypothesis management, where as NLG is about choice; When producing NL utterances from non-linguistic material, everything bit of information to be encoded in the output has to be chosen along the way.

14/50

slide-35
SLIDE 35

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Comparison to NLU

Differences

NLU is about hypothesis management, where as NLG is about choice; When producing NL utterances from non-linguistic material, everything bit of information to be encoded in the output has to be chosen along the way. parsing, or most shallow NLU tasks, ignore certain aspects of meaning; NLG utilizes elaborated meaning representations

14/50

slide-36
SLIDE 36

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Comparison to NLU

Differences

NLU is about hypothesis management, where as NLG is about choice; When producing NL utterances from non-linguistic material, everything bit of information to be encoded in the output has to be chosen along the way. parsing, or most shallow NLU tasks, ignore certain aspects of meaning; NLG utilizes elaborated meaning representations research in NLG has focused more on texts than has NLU

14/50

slide-37
SLIDE 37

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Texts

For NLG, we often need to deal with structures larger than the sentence.

15/50

slide-38
SLIDE 38

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Texts

For NLG, we often need to deal with structures larger than the sentence.

Definition

Text: A unit of language larger than the individual clause (spoken or written). This is referred to as a discourse in J&M.

15/50

slide-39
SLIDE 39

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Texts

For NLG, we often need to deal with structures larger than the sentence.

Definition

Text: A unit of language larger than the individual clause (spoken or written). This is referred to as a discourse in J&M. a phone conversation

15/50

slide-40
SLIDE 40

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Texts

For NLG, we often need to deal with structures larger than the sentence.

Definition

Text: A unit of language larger than the individual clause (spoken or written). This is referred to as a discourse in J&M. a phone conversation a recipe

15/50

slide-41
SLIDE 41

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Texts

For NLG, we often need to deal with structures larger than the sentence.

Definition

Text: A unit of language larger than the individual clause (spoken or written). This is referred to as a discourse in J&M. a phone conversation a recipe a weather report

15/50

slide-42
SLIDE 42

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Texts

For NLG, we often need to deal with structures larger than the sentence.

Definition

Text: A unit of language larger than the individual clause (spoken or written). This is referred to as a discourse in J&M. a phone conversation a recipe a weather report a paragraph in War and Peace

15/50

slide-43
SLIDE 43

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Texts

For NLG, we often need to deal with structures larger than the sentence.

Definition

Text: A unit of language larger than the individual clause (spoken or written). This is referred to as a discourse in J&M. a phone conversation a recipe a weather report a paragraph in War and Peace an entire novel

15/50

slide-44
SLIDE 44

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Text coherence

Such texts exhibit a certain structure and we can speak of their well-formedness, just like any other unit of language.

16/50

slide-45
SLIDE 45

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Text coherence

Such texts exhibit a certain structure and we can speak of their well-formedness, just like any other unit of language.

Definition

A coherent text is one whose parts are interrelated in meaningful way. An incoherent text is one whose parts do not bind together in a naturalistic manner.

16/50

slide-46
SLIDE 46

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Text coherence

Such texts exhibit a certain structure and we can speak of their well-formedness, just like any other unit of language.

Definition

A coherent text is one whose parts are interrelated in meaningful way. An incoherent text is one whose parts do not bind together in a naturalistic manner.

16/50

slide-47
SLIDE 47

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Text coherence

Such texts exhibit a certain structure and we can speak of their well-formedness, just like any other unit of language.

Definition

A coherent text is one whose parts are interrelated in meaningful way. An incoherent text is one whose parts do not bind together in a naturalistic manner. Pronominalization

16/50

slide-48
SLIDE 48

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Text coherence

Such texts exhibit a certain structure and we can speak of their well-formedness, just like any other unit of language.

Definition

A coherent text is one whose parts are interrelated in meaningful way. An incoherent text is one whose parts do not bind together in a naturalistic manner. Pronominalization Temporal expressions

16/50

slide-49
SLIDE 49

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Text coherence: pronominal expressions

James Riddle “Jimmy” Hoffa (born February 14, 1913, disappeared July 30, 1975), was an American labor leader. As the president of the International Brotherhood of Teamsters from the mid-1950s to the mid-1960s, Hoffa wielded considerable influence. After he was convicted of attempted bribery of a grand juror, he served nearly a decade in prison. He is also well-known in popular culture for the mysterious circumstances surrounding his unexplained disappearance and presumed death. His son James P. Hoffa is the current president of the Teamsters.

17/50

slide-50
SLIDE 50

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Text coherence: pronominal expressions

James Riddle “Jimmy” Hoffa (born February 14, 1913, disappeared July 30, 1975), was an American labor leader. As the president of the International Brotherhood of Teamsters from the mid-1950s to the mid-1960s, Hoffa wielded considerable influence. After he was convicted of attempted bribery of a grand juror, he served nearly a decade in prison. He is also well-known in popular culture for the mysterious circumstances surrounding his unexplained disappearance and presumed death. His son James P. Hoffa is the current president of the Teamsters.

18/50

slide-51
SLIDE 51

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Text coherence: temporal expressions

During Buck’s time in jail, Clyde had been the driver in a store robbery. The wife of the murder victim, when shown photos, picked Clyde as one of the shooters. On August 5, 1932, while Bonnie was visiting her mother, Clyde and two associates were drinking alcohol at a dance in Stringtown, Oklahoma (illegal under Prohibition). When they were approached by sheriff C.G. Maxwell and his deputy, Clyde

  • pened fire, killing deputy Eugene C. Moore. That was the

first killing of a lawman by what was later known as the Barrow Gang, a total which would eventually amount to nine slain officers.

19/50

slide-52
SLIDE 52

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Text coherence: temporal expressions

During Buck’s time in jail, Clyde had been the driver in a store robbery. The wife of the murder victim, when shown photos, picked Clyde as one of the shooters. On August 5, 1932, while Bonnie was visiting her mother, Clyde and two associates were drinking alcohol at a dance in Stringtown, Oklahoma (illegal under Prohibition). When they were approached by sheriff C.G. Maxwell and his deputy, Clyde

  • pened fire, killing deputy Eugene C. Moore. That was the

first killing of a lawman by what was later known as the Barrow Gang, a total which would eventually amount to nine slain officers.

20/50

slide-53
SLIDE 53

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Evaluation

Problem: evaluation of NLG systems is always a great challenge (no gold standard)

21/50

slide-54
SLIDE 54

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Evaluation

Problem: evaluation of NLG systems is always a great challenge (no gold standard) Evaluation techniques:

21/50

slide-55
SLIDE 55

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Evaluation

Problem: evaluation of NLG systems is always a great challenge (no gold standard) Evaluation techniques: Turing Test (subjective)

21/50

slide-56
SLIDE 56

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Evaluation

Problem: evaluation of NLG systems is always a great challenge (no gold standard) Evaluation techniques: Turing Test (subjective) task-oriented (expensive)

21/50

slide-57
SLIDE 57

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Evaluation

Problem: evaluation of NLG systems is always a great challenge (no gold standard) Evaluation techniques: Turing Test (subjective) task-oriented (expensive) statistical comparison with real texts (untrustworthy)

21/50

slide-58
SLIDE 58

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Turing Test for evaluation of an NLG system

5 Indistinguishable from human 4 Most likely human 3 Maybe human or machine 2 Most likely machine 1 Definitely machine

22/50

slide-59
SLIDE 59

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Grice’s Conversational Maxims

23/50

slide-60
SLIDE 60

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Grice’s Conversational Maxims

Quantity: Make your contribution as informative as is required

23/50

slide-61
SLIDE 61

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Grice’s Conversational Maxims

Quantity: Make your contribution as informative as is required Quality: Do not say what you believe is false. Do not say that for which you lack adequate evidence.

23/50

slide-62
SLIDE 62

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Grice’s Conversational Maxims

Quantity: Make your contribution as informative as is required Quality: Do not say what you believe is false. Do not say that for which you lack adequate evidence. Relation: Be relevant.

23/50

slide-63
SLIDE 63

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Grice’s Conversational Maxims

Quantity: Make your contribution as informative as is required Quality: Do not say what you believe is false. Do not say that for which you lack adequate evidence. Relation: Be relevant. Manner: Be perspicuous, avoid obscurity of expression, avoid ambiguity, be brief, be orderly.

23/50

slide-64
SLIDE 64

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Statistical evaluation scenarios

Comparison with humanly produced texts:

24/50

slide-65
SLIDE 65

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Statistical evaluation scenarios

Comparison with humanly produced texts: text length, mean length of utterance (MLU)

24/50

slide-66
SLIDE 66

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Statistical evaluation scenarios

Comparison with humanly produced texts: text length, mean length of utterance (MLU) average number of embedded clauses (and other complex structures)

24/50

slide-67
SLIDE 67

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Statistical evaluation scenarios

Comparison with humanly produced texts: text length, mean length of utterance (MLU) average number of embedded clauses (and other complex structures) diction (number of word types)

24/50

slide-68
SLIDE 68

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Statistical evaluation scenarios

Comparison with humanly produced texts: text length, mean length of utterance (MLU) average number of embedded clauses (and other complex structures) diction (number of word types) number of long distance dependencies

24/50

slide-69
SLIDE 69

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

General evaluation criteria

Despite the challenges, we can posit some fundamental criteria for evaluating NLG systems:

25/50

slide-70
SLIDE 70

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

General evaluation criteria

Despite the challenges, we can posit some fundamental criteria for evaluating NLG systems: content: does the output contain appropriate/enough information?

25/50

slide-71
SLIDE 71

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

General evaluation criteria

Despite the challenges, we can posit some fundamental criteria for evaluating NLG systems: content: does the output contain appropriate/enough information?

  • rganization: is the discourse structure realistic?

25/50

slide-72
SLIDE 72

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

General evaluation criteria

Despite the challenges, we can posit some fundamental criteria for evaluating NLG systems: content: does the output contain appropriate/enough information?

  • rganization: is the discourse structure realistic?

correctness: are there grammatical or stylistic errors?

25/50

slide-73
SLIDE 73

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

General evaluation criteria

Despite the challenges, we can posit some fundamental criteria for evaluating NLG systems: content: does the output contain appropriate/enough information?

  • rganization: is the discourse structure realistic?

correctness: are there grammatical or stylistic errors? textual flow: is the language choppy or smooth?

25/50

slide-74
SLIDE 74

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG subtask: Non-linguistic

There are several tasks for a full generation system:

26/50

slide-75
SLIDE 75

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG subtask: Non-linguistic

There are several tasks for a full generation system: content determination: task of deciding what information is to be communicated

26/50

slide-76
SLIDE 76

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG subtask: Non-linguistic

There are several tasks for a full generation system: content determination: task of deciding what information is to be communicated discourse structuring: deciding how to package the ‘chunks’ of content

26/50

slide-77
SLIDE 77

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG subtask: Linguistic

27/50

slide-78
SLIDE 78

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG subtask: Linguistic

lexicalization: determine the particular words and construction types to use

27/50

slide-79
SLIDE 79

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG subtask: Linguistic

lexicalization: determine the particular words and construction types to use aggregation: decide how much information to include in each sentence/phrase

27/50

slide-80
SLIDE 80

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG subtask: Linguistic

lexicalization: determine the particular words and construction types to use aggregation: decide how much information to include in each sentence/phrase referring expression generation: determine what expressions should be used to refer to the entities in the domain

27/50

slide-81
SLIDE 81

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG subtask: Linguistic

lexicalization: determine the particular words and construction types to use aggregation: decide how much information to include in each sentence/phrase referring expression generation: determine what expressions should be used to refer to the entities in the domain linguistic realization: task of converting abstract representations of sentences to real text; linearization.

27/50

slide-82
SLIDE 82

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG subtask: Linguistic

lexicalization: determine the particular words and construction types to use aggregation: decide how much information to include in each sentence/phrase referring expression generation: determine what expressions should be used to refer to the entities in the domain linguistic realization: task of converting abstract representations of sentences to real text; linearization. structure realization: task of converting abstract structures such as paragraphs and sections into markup symbols understood by the document presentation component (e.g., HTML, L

AT

EX)

27/50

slide-83
SLIDE 83

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG subtask: Linguistic

lexicalization: determine the particular words and construction types to use aggregation: decide how much information to include in each sentence/phrase referring expression generation: determine what expressions should be used to refer to the entities in the domain linguistic realization: task of converting abstract representations of sentences to real text; linearization. structure realization: task of converting abstract structures such as paragraphs and sections into markup symbols understood by the document presentation component (e.g., HTML, L

AT

EX) Need to modularize the above tasks appropriately, given the goal of the NLG system and the available resources.

27/50

slide-84
SLIDE 84

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Lexicalization

(1) a. Mary’s car

  • b. the car owned by Mary

28/50

slide-85
SLIDE 85

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Lexicalization

(3) a. Mary’s car

  • b. the car owned by Mary

(4) a. the ship’s cargo hold

  • b. the cargo hold which is part of the ship

28/50

slide-86
SLIDE 86

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Agrregation

(5) a. I am Ron Paul. I am a rogue Republican.

  • b. My name is Ron Paul and I’m a rogue Republican.

29/50

slide-87
SLIDE 87

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Agrregation

(7) a. I am Ron Paul. I am a rogue Republican.

  • b. My name is Ron Paul and I’m a rogue Republican.

(8) a. The course number is ling571.

  • b. The course is difficult. The course is open to

undergraduates.

  • c. Ling571 is difficult, but open to undergraduates.

29/50

slide-88
SLIDE 88

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Agrregation: other examples

30/50

slide-89
SLIDE 89

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Agrregation: other examples

John’s bicycle is red

30/50

slide-90
SLIDE 90

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Agrregation: other examples

John’s bicycle is red Mary’s bicycle is yellow

30/50

slide-91
SLIDE 91

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Agrregation: other examples

John’s bicycle is red Mary’s bicycle is yellow Tom’s bicycle is blue

30/50

slide-92
SLIDE 92

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Agrregation: other examples

John’s bicycle is red Mary’s bicycle is yellow Tom’s bicycle is blue Lisa’s bicycle is red

30/50

slide-93
SLIDE 93

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Agrregation: other examples

John’s bicycle is red Mary’s bicycle is yellow Tom’s bicycle is blue Lisa’s bicycle is red becomes:

30/50

slide-94
SLIDE 94

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Agrregation: other examples

John’s bicycle is red Mary’s bicycle is yellow Tom’s bicycle is blue Lisa’s bicycle is red becomes: John and Lisa have red bicycles.

30/50

slide-95
SLIDE 95

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Agrregation: other examples

John’s bicycle is red Mary’s bicycle is yellow Tom’s bicycle is blue Lisa’s bicycle is red becomes: John and Lisa have red bicycles. Tom’s and Mary’s bicycles are blue and yellow respectively.

30/50

slide-96
SLIDE 96

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Lexical aggregation

31/50

slide-97
SLIDE 97

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Lexical aggregation

Ericsson made profit in 2004

31/50

slide-98
SLIDE 98

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Lexical aggregation

Ericsson made profit in 2004 Nokia made profit in 2004

31/50

slide-99
SLIDE 99

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Lexical aggregation

Ericsson made profit in 2004 Nokia made profit in 2004 Siemens made profit in 2004

31/50

slide-100
SLIDE 100

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Lexical aggregation

Ericsson made profit in 2004 Nokia made profit in 2004 Siemens made profit in 2004 ATT made profit in 2004

31/50

slide-101
SLIDE 101

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Lexical aggregation

Ericsson made profit in 2004 Nokia made profit in 2004 Siemens made profit in 2004 ATT made profit in 2004 becomes:

31/50

slide-102
SLIDE 102

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Lexical aggregation

Ericsson made profit in 2004 Nokia made profit in 2004 Siemens made profit in 2004 ATT made profit in 2004 becomes: All telecom companies in the world except Alcatel made profit in 2004

31/50

slide-103
SLIDE 103

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Referring Expressions

(9) a. I know that guy.

  • b. I know Noam Chomsky.

32/50

slide-104
SLIDE 104

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Referring Expressions

(11) a. I know that guy.

  • b. I know Noam Chomsky.

(12) a. It’s a long-legged, hairy one.

  • b. The European wolf spider is a long-legged, hairy

spider.

32/50

slide-105
SLIDE 105

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Today’s lecture

1

Demos

2

Basics of NLG NLG concepts Issues in NLG NLG subtasks

3

Architecture of NLG systems Two-step architecture Three-step architecture

4

Hw7

33/50

slide-106
SLIDE 106

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Ways to generate

A generation system can be devised for individual utterances

  • r whole texts.

34/50

slide-107
SLIDE 107

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Ways to generate

A generation system can be devised for individual utterances

  • r whole texts.

canned text (easy, brittle)

34/50

slide-108
SLIDE 108

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Ways to generate

A generation system can be devised for individual utterances

  • r whole texts.

canned text (easy, brittle) template-based generation (easy, more flexible, ad hoc)

34/50

slide-109
SLIDE 109

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Ways to generate

A generation system can be devised for individual utterances

  • r whole texts.

canned text (easy, brittle) template-based generation (easy, more flexible, ad hoc) feature collection and linearization (very hard, flexible, theory driven)

34/50

slide-110
SLIDE 110

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Templates from ELIZA program

( "Perhaps you don’t want to %1.", "Do you want to be able to %1?", "If you could %1, would you?")), ... ( "Why do you think I am %1?", "Does it please you to think that I’m %1?", "Perhaps you would like me to be %1.", "Perhaps you’re really talking about yourself?")),

35/50

slide-111
SLIDE 111

NLP reference architecture

slide-112
SLIDE 112

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

NLG architectures paradigms

Two common architectures (two- and three- step systems) Based on splitting up the NLG subtasks (e.g., textplanner, lexical aggregation). Based on how complex the system needs to be.

37/50

slide-113
SLIDE 113

NLG architecture: Two-step

slide-114
SLIDE 114

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Two-step: Main components

1 Deep generation: determines and structures content

  • f resulting text; insert words (lexicalization); map

message to linguistic structure

2 Surface generation: fill in grammatical details, some

lexicalization

39/50

slide-115
SLIDE 115

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Two-step: Example systems

Used in simpler, early systems:

FOG System: generates weather reports from numerical weather simulations. Peba: generates taxonomic descriptions or comparisons

  • f animals from a knowledge base of animal facts.

Problems in modularization and control over choice. In general, it’s really hard to strictly keep the non-linguistic choices separate from the linguistic choices.

40/50

slide-116
SLIDE 116

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Example from FOG System: weather text

Winds southwest 15 diminishing to light, late this evening. Winds light Friday. Showers ending late this evening. Fog. Outlook for Saturday...light winds.

41/50

slide-117
SLIDE 117

PEBA system: knowledge base sample

(hasprop Echidna (linean-classification Family)) (distinguishing-characteristic Echidna Monotreme (body-covering sharp-spines)) (hasprop Echidna (nose long-snout)) (hasprop Echidna (social-living-status lives-by-itself)) (hasprop Echidna (diet eats-ants-termites-earthworms)) (hasprop Echidna (activity-time active-at-dusk-dawn)) (hasprop Echidna (colouring browny-black-coat-paler-coloured-spines)) (hasprop Echidna (lifespan lifespan-50-years-captivity)

slide-118
SLIDE 118

PEBA system: knowledge base sample

(hasprop Echidna (linean-classification Family)) (distinguishing-characteristic Echidna Monotreme (body-covering sharp-spines)) (hasprop Echidna (nose long-snout)) (hasprop Echidna (social-living-status lives-by-itself)) (hasprop Echidna (diet eats-ants-termites-earthworms)) (hasprop Echidna (activity-time active-at-dusk-dawn)) (hasprop Echidna (colouring browny-black-coat-paler-coloured-spines)) (hasprop Echidna (lifespan lifespan-50-years-captivity)

text

The small spined monotreme belongs to the Echidna Family. Its nose is a long snout. The Echidna lives by itself.

slide-119
SLIDE 119

NLG architecture: Three-step

slide-120
SLIDE 120

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Three-step: Example systems

Better systems

More flexible, more modular. More control over the output. WeatherReporter: more complex, cohesive weather reports KNIGHT System: a biology explanation system, from knowledge base to explanatory paragraphs More flexible, more modular. More control over the output.

44/50

slide-121
SLIDE 121

Three-step: Main components

Module Content task Structure task Text Planner content determi- nation rhetorical structuring Microplanner lexicalization; re- ferring expression generation aggregation Surface Realizer linguistic realiza- tion structure realization

slide-122
SLIDE 122

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

WeatherReporter: weather text

The month was rather dry with only three days of rain in the middle of the month. The total for the year so far is very depleted again.

46/50

slide-123
SLIDE 123

KNIGHT System KB

slide-124
SLIDE 124

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

KNIGHT System Output

Embryo sac formation is a kind of female gametophyte

  • formation. During embryo sac formation, the embryo sac is

formed from the megaspore mother cell. Embryo sac formation occurs in the ovule.

48/50

slide-125
SLIDE 125

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Today’s lecture

1

Demos

2

Basics of NLG NLG concepts Issues in NLG NLG subtasks

3

Architecture of NLG systems Two-step architecture Three-step architecture

4

Hw7

49/50

slide-126
SLIDE 126

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Hw7, main points

Tasks

50/50

slide-127
SLIDE 127

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Hw7, main points

Tasks

1 Create code to process an input text specification

(XML)

50/50

slide-128
SLIDE 128

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Hw7, main points

Tasks

1 Create code to process an input text specification

(XML)

2 Build a microplanner (in language of your choice) 50/50

slide-129
SLIDE 129

Natural Language Generation Scott Farrar CLMA, University

  • f Washington far-

rar@u.washington.edu Demos Basics of NLG

NLG concepts Issues in NLG NLG subtasks

Architecture of NLG systems

Two-step architecture Three-step architecture

Hw7

Hw7, main points

Tasks

1 Create code to process an input text specification

(XML)

2 Build a microplanner (in language of your choice) 3 Use the SimpleNLG (v.4) surface realizer (Java) 50/50