natural language generation
play

Natural Language Generation . .. . . .. .. . .. . . . .. - PowerPoint PPT Presentation

. .. . . .. . . .. . . .. . . .. . . . . Ondej Duek Ondej Duek 1/ 40 May 14 th , 2015 Charles University in Prague Faculty of Mathematics and Physics Institute of Formal and Applied Linguistics for Spoken Dialogue


  1. • Paiva&Evans : linguistic features annotated in corpus generated with • PERSONAGE-PE : personality traits connected to linguistic features via . . . .. . . .. .. . .. . . .. . . . . . . .. . . .. . Example NLG Systems Sentence planning Trainable Sentence Planning: Parameter Optimization Examples many parameter settings, correlation analysis machine learning 9/ 40 Ondřej Dušek .. .. . . . . .. . . .. . . .. . . .. . . .. .. . .. . . .. . . .. . Natural Language Generation . .. . . .. . . • Requires a flexible handcrafed planner • No overgeneration • Adjusting its parameters “somehow”

  2. . . .. . . .. . . .. . . .. . . .. . .. . Sentence planning Ondřej Dušek 9/ 40 machine learning many parameter settings, correlation analysis Examples Trainable Sentence Planning: Parameter Optimization Example NLG Systems . . .. . . .. . . .. .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . .. . Natural Language Generation . .. . . .. . . .. • Requires a flexible handcrafed planner • No overgeneration • Adjusting its parameters “somehow” • Paiva&Evans : linguistic features annotated in corpus generated with • PERSONAGE-PE : personality traits connected to linguistic features via

  3. • General purpose • Functional Unification . .. . . .. . . . . . .. .. . .. . .. .. . KPML Ondřej Dušek 10/ 40 Grammar FUF/SURGE Grammar multilingual Grammar-based Realizers (90's): KPML , FUF/SURGE .. Surface Realization Example NLG Systems . .. . . . . . . . . .. . . .. . . .. . . .. . . .. . . . .. . . .. . . .. .. Natural Language Generation . . .. . . .. . (EXAMPLE :NAME EX-SET-1 :TARGETFORM "It is raining cats and dogs." :LOGICALFORM • General purpose, (A / AMBIENT-PROCESS :LEX RAIN :TENSE PRESENT-CONTINUOUS :ACTEE (C / OBJECT :LEX CATS-AND-DOGS :NUMBER MASS)) ) • Systemic Functional

  4. . . . . .. . . .. . . .. .. . .. . . .. . . KPML Ondřej Dušek 10/ 40 Grammar FUF/SURGE Grammar multilingual Grammar-based Realizers (90's): KPML , FUF/SURGE .. Surface Realization Example NLG Systems . .. . . .. . . . . . .. . .. .. . . .. . . .. . . .. . . .. . . .. .. . . Natural Language Generation . . . .. . . . .. (EXAMPLE :NAME EX-SET-1 :TARGETFORM "It is raining cats and dogs." :LOGICALFORM • General purpose, (A / AMBIENT-PROCESS :LEX RAIN :TENSE PRESENT-CONTINUOUS :ACTEE (C / OBJECT :LEX CATS-AND-DOGS :NUMBER MASS)) ) • Systemic Functional • General purpose • Functional Unification

  5. . . .. . . .. . . .. . . .. . . .. . .. . Surface Realization Ondřej Dušek 11/ 40 enhancements Grammar multi-lingual Grammar-based Realizer: OpenCCG Example NLG Systems . . .. . . .. . .. . .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . .. . . .. . . .. . . .. Natural Language Generation • General purpose, • Combinatory Categorial • Used in several projects • With statistical

  6. . . .. . . .. .. . .. . . .. . . .. . .. . . . .. . . .. . Example NLG Systems Surface Realization Procedural Realizer: SimpleNLG other languages (procedural) 12/ 40 Ondřej Dušek . . .. .. . . .. . . .. . . . . . .. . . .. . . . .. . . . .. Natural Language Generation . . . . .. . .. .. Lexicon lexicon = new XMLLexicon("my-lexicon.xml"); • General purpose NLGFactory nlgFactory = new NLGFactory(lexicon); Realiser realiser = new Realiser(lexicon); • English, adapted to several SPhraseSpec p = nlgFactory.createClause(); p.setSubject("Mary"); p.setVerb("chase"); • Java implementation p.setObject("the monkey"); p.setFeature(Feature.TENSE, Tense.PAST); String output = realiser.realiseSentence(p); System.out.println(output); >>> Mary chased the monkey.

  7. • Ranking according to: • n -gram models ( NITROGEN, HALOGEN ) • Tree models (XTAG grammar – FERGUS ) • Predicted Text-to-Speech quality ( Nakatsu and White ) • Personality traits (extraversion, agreeableness… – CRAG ) • Provides variance, but at a greater computational cost . .. . . .. . . .. . . . .. . . .. .. . . .. . . .. . Example NLG Systems Surface Realization Trainable Realizers: Overgenerate and Rank + alignment (repeating words uttered by dialogue counterpart) 13/ 40 Ondřej Dušek .. . . . . . .. . . .. . . .. . . .. . . . .. . . .. . . .. . . .. . Natural Language Generation .. .. . . .. . . • Require a handcrafued realizer, e.g. CCG realizer • Input underspecified → more outputs possible • Overgenerate • Then use a statistical reranker

  8. • Provides variance, but at a greater computational cost . . .. .. . .. . . .. . . .. . . .. .. . . . . .. . . .. . Example NLG Systems Surface Realization Trainable Realizers: Overgenerate and Rank + alignment (repeating words uttered by dialogue counterpart) 13/ 40 Ondřej Dušek . . .. .. . . .. . . .. . . .. . . . . . .. . . . .. . . . .. Natural Language Generation . .. . . .. . . .. • Require a handcrafued realizer, e.g. CCG realizer • Input underspecified → more outputs possible • Overgenerate • Then use a statistical reranker • Ranking according to: • n -gram models ( NITROGEN, HALOGEN ) • Tree models (XTAG grammar – FERGUS ) • Predicted Text-to-Speech quality ( Nakatsu and White ) • Personality traits (extraversion, agreeableness… – CRAG )

  9. . . . .. .. . .. . . .. . . .. . . .. . .. .. . . .. . . .. . Example NLG Systems Surface Realization Trainable Realizers: Overgenerate and Rank + alignment (repeating words uttered by dialogue counterpart) 13/ 40 Ondřej Dušek . . . .. . . .. . . .. . . .. . . .. . . . . . .. . . . .. . Natural Language Generation .. . . .. .. . . • Require a handcrafued realizer, e.g. CCG realizer • Input underspecified → more outputs possible • Overgenerate • Then use a statistical reranker • Ranking according to: • n -gram models ( NITROGEN, HALOGEN ) • Tree models (XTAG grammar – FERGUS ) • Predicted Text-to-Speech quality ( Nakatsu and White ) • Personality traits (extraversion, agreeableness… – CRAG ) • Provides variance, but at a greater computational cost

  10. . .. .. .. . . .. . . .. . . .. . . . .. . .. . . .. . . .. . Example NLG Systems Surface Realization Trainable Realizers: Syntax-Based 14/ 40 Ondřej Dušek . . . .. . . .. . . .. . . .. . . .. . . . . . .. . . . . .. .. . . .. . . .. Natural Language Generation • StuMaBa : general realizer based on SVMs • Pipeline: ↓ Deep syntax/semantics ↓ surface syntax ↓ linearization ↓ morphologization

  11. • Most common, also in commercial NLG systems • Simple, straightforward, reliable (custom-tailored for domain) • Lack generality and variation, difficult to maintain • Enhancements for more complex utterances: rules . . . .. . . .. . . .. .. . .. . . .. . . .. . . .. . Example NLG Systems Holistic NLG Approaches Holistic NLG Holistic NLG Template-based systems 15/ 40 Ondřej Dušek . .. . . . . .. . . .. . . .. . . .. . . .. . . .. .. . . .. . . .. Natural Language Generation . . .. . . .. . . • Only one stage – no distinction • “Good enough” for limited domains, also in SDS

  12. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . Example NLG Systems Holistic NLG Approaches Holistic NLG Holistic NLG Template-based systems 15/ 40 Ondřej Dušek . .. .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . . . .. .. . . .. . . .. Natural Language Generation • Only one stage – no distinction • “Good enough” for limited domains, also in SDS • Most common, also in commercial NLG systems • Simple, straightforward, reliable (custom-tailored for domain) • Lack generality and variation, difficult to maintain • Enhancements for more complex utterances: rules

  13. . . .. . . .. . . .. . . .. . . .. . .. . Holistic NLG Approaches Ondřej Dušek 16/ 40 Facebook templates domain) Alex (English restaurant Example: Templates Example NLG Systems . . .. . . .. . . .. .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . .. . Natural Language Generation . . . .. . . .. .. • Just filling variables into slots • Possibly a few enhancements, e. g. articles inform(pricerange="{pricerange}"): 'It is in the {pricerange} price range.' affirm()&inform(task="find") &inform(pricerange="{pricerange}"): 'Ok, you are looking for something in the' + ' {pricerange} price range.' affirm()&inform(area="{area}"): 'Ok, you want something in the {area} area.' affirm()&inform(food="{food}") &inform(pricerange="{pricerange}"): 'Ok, you want something with the {food} food' + ' in the {pricerange} price range.' inform(food="None"): 'I do not have any information' + ' about the type of food.'

  14. • BAGEL : Bayesian networks • semantic stacks, ordering • Angeli et al. : log-linear model • records • WASP • noisy channel, similar to MT .. . . .. . . . .. . . . .. . . . .. .. . . . . Example NLG Systems Holistic NLG Approaches Statistical Holistic NLG (typically: MR + sentence + alignment) Examples fields templates 1 : Synchronous CFGs 17/ 40 Ondřej Dušek .. .. . .. .. . . .. . . . .. . .. . . .. . . . . .. .. . . .. . . .. . . . . . .. . . .. . Natural Language Generation • Limited domain • Based on supervised learning • Typically: phrase-based

  15. . . .. .. . .. . . .. . . .. . . .. . .. . . . .. . . .. . Example NLG Systems Holistic NLG Approaches Statistical Holistic NLG (typically: MR + sentence + alignment) Examples 17/ 40 Ondřej Dušek . . .. .. . . .. . . .. . . .. . . . . . .. . . . .. . . . .. Natural Language Generation . . . . .. .. . .. • Limited domain • Based on supervised learning • Typically: phrase-based • BAGEL : Bayesian networks • semantic stacks, ordering • Angeli et al. : log-linear model • records ↘ fields ↘ templates • WASP − 1 : Synchronous CFGs • noisy channel, similar to MT

  16. • Our generator learns alignments jointly • (with sentence planning) • training from pairs: MR + sentence . .. . . .. . . .. . . . . . .. .. .. .. . .. . . .. . Our System Overview Our experiments: Two-Step NLG for SDS Learning from unaligned data a) requires detailed alignments of MR elements and words/phrases b) uses a separate alignment step 18/ 40 Ondřej Dušek . . . . . . .. . . .. . . .. . . .. . . .. . . . .. . . .. . .. .. . . .. . . .. . . Natural Language Generation • Typical NLG training: MR inform(name=X, type=placetoeat, eattype=restaurant, area=riverside, food=Italian) alignment X is an italian restaurant in the riverside area . text

  17. . .. . .. .. . . .. . . .. . . .. . . . . Our experiments: Two-Step NLG for SDS Ondřej Dušek 18/ 40 b) uses a separate alignment step of MR elements and words/phrases a) requires detailed alignments Learning from unaligned data Overview . Our System . .. . . .. .. . . . . .. .. . . .. . . .. . . .. . . .. . . . . . .. . .. .. . . .. . . .. . . Natural Language Generation • Typical NLG training: • Our generator learns alignments jointly • (with sentence planning) • training from pairs: MR + sentence MR inform(name=X, type=placetoeat, eattype=restaurant, area=riverside, food=Italian) X is an italian restaurant in the riverside area . text

  18. • Step 1. – sentence planning • statistical, our main focus • Sentence plan : deep-syntax dependency trees • based on TectoMT 's t-layer, but very simplified • two attributes per tree node: t-lemma + formeme • using surface word order • Step 2. – surface realization • reusing Treex / TectoMT English synthesis (rule-based) • Output : plain text sentence . .. . . .. . . .. . . . .. . . .. . . .. . . .. . Our System Workflow / data formats Overall workflow of our generator 19/ 40 Ondřej Dušek .. . .. .. .. . . .. . . . . . .. . . .. . . . . .. . . .. . . .. . . .. . . .. . . .. . Natural Language Generation • Input : a MR • here – dialogue acts: “inform” + slot-value pairs • other formats possible

  19. • Sentence plan : deep-syntax dependency trees • based on TectoMT 's t-layer, but very simplified • two attributes per tree node: t-lemma + formeme • using surface word order • Step 2. – surface realization • reusing Treex / TectoMT English synthesis (rule-based) • Output : plain text sentence . .. . . .. . . .. . . . .. .. . .. .. . . .. . . .. . Our System Workflow / data formats Overall workflow of our generator 19/ 40 Ondřej Dušek . . . . . . .. . . .. . . .. . . .. . . . .. . . .. . . .. . . .. . Natural Language Generation .. .. . . .. . . • Input : a MR • here – dialogue acts: “inform” + slot-value pairs • other formats possible • Step 1. – sentence planning • statistical, our main focus

  20. • Output : plain text sentence . .. . .. . .. .. . . .. . . .. . . . . .. .. . . .. . . .. . Our System Workflow / data formats Overall workflow of our generator 19/ 40 Ondřej Dušek . . . .. . . .. . . .. . . .. . . .. . . . . . .. . . . .. . Natural Language Generation .. . . .. .. . . • Input : a MR • here – dialogue acts: “inform” + slot-value pairs • other formats possible • Step 1. – sentence planning • statistical, our main focus • Sentence plan : deep-syntax dependency trees • based on TectoMT 's t-layer, but very simplified • two attributes per tree node: t-lemma + formeme • using surface word order • Step 2. – surface realization • reusing Treex / TectoMT English synthesis (rule-based)

  21. . .. . .. . .. .. . . .. . . .. . . . .. . .. . . .. . . .. . Our System Workflow / data formats Overall workflow of our generator 19/ 40 Ondřej Dušek . . . .. . . .. . . .. . . .. . . .. . . . . . .. . . . .. . Natural Language Generation .. . . .. . .. . • Input : a MR • here – dialogue acts: “inform” + slot-value pairs • other formats possible • Step 1. – sentence planning • statistical, our main focus • Sentence plan : deep-syntax dependency trees • based on TectoMT 's t-layer, but very simplified • two attributes per tree node: t-lemma + formeme • using surface word order • Step 2. – surface realization • reusing Treex / TectoMT English synthesis (rule-based) • Output : plain text sentence

  22. . .. . .. . . .. .. . .. . . .. . . . .. . .. . . .. . . .. . Our System Workflow / data formats Data structures used 20/ 40 Ondřej Dušek . . . .. . . .. . . .. . . .. . . .. . . . . . .. . . . .. . Natural Language Generation .. . . .. .. . . inform(name=X, type=placetoeat, eattype=restaurant, area=riverside, food=Italian) t-tree be v:fin X-name restaurant n:subj n:obj italian area adj:attr n:in+X riverside n:attr X is an italian restaurant in the riverside area .

  23. • The 2nd step – rule based – can ensure grammatical • or at least it's more straightforward to fix when it doesn't • The realizer is (relatively) easy to implement and • + why not use it if we have it already in Treex / TectoMT . . .. . . .. . . .. .. . . .. . . .. . . .. . . .. . Our System Two-step architecture Why we keep the two-step approach correctness domain-independent 21/ 40 Ondřej Dušek . .. . . . . .. . . .. . . .. . . .. . . .. . . . .. . . .. . . .. Natural Language Generation .. . .. . . .. . . • It makes the 1st – statistical – task simpler • no need to worry about morphology • this will be more important for Czech (and similar)

  24. • The realizer is (relatively) easy to implement and • + why not use it if we have it already in Treex / TectoMT . . .. . . .. . . .. . . .. . . . . .. . . .. . . .. . Our System Two-step architecture Why we keep the two-step approach correctness domain-independent 21/ 40 Ondřej Dušek .. .. . . .. . .. . . .. . . .. . . .. . . .. . . .. . . .. . . .. Natural Language Generation . . .. . . .. . . • It makes the 1st – statistical – task simpler • no need to worry about morphology • this will be more important for Czech (and similar) • The 2nd step – rule based – can ensure grammatical • or at least it's more straightforward to fix when it doesn't

  25. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . Our System Two-step architecture Why we keep the two-step approach correctness domain-independent 21/ 40 Ondřej Dušek . .. .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . . .. Natural Language Generation . .. . . .. . . .. • It makes the 1st – statistical – task simpler • no need to worry about morphology • this will be more important for Czech (and similar) • The 2nd step – rule based – can ensure grammatical • or at least it's more straightforward to fix when it doesn't • The realizer is (relatively) easy to implement and • + why not use it if we have it already in Treex / TectoMT

  26. • but we can do it easily using Treex • automatic annotation is good enough . .. . . .. . . . . . .. . . .. .. .. . .. . . .. . . .. . Our System Two-step architecture Downside of the two-step approach – t-layer analysis implemented for several languages 22/ 40 Ondřej Dušek . . .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . .. . . .. . . .. . . .. Natural Language Generation • We need to analyze training sentences into deep trees

  27. . . . .. . . .. . . .. . . .. . . .. . .. .. . . .. . . .. . Our System Two-step architecture Downside of the two-step approach – t-layer analysis implemented for several languages 22/ 40 Ondřej Dušek .. . . .. . . .. . . .. . . .. . . .. . . . . . .. . . .. . . .. . . .. . . .. Natural Language Generation • We need to analyze training sentences into deep trees • but we can do it easily using Treex • automatic annotation is good enough

  28. • A*-style search • incrementally finding the path • from an empty tree • to a full sentence plan tree which contains all information • using open_set , close_set – heaps sorted by score . . .. . . .. . . .. . . .. . . .. .. . . .. . . .. . Our System Sentence planner Sentence planner – overall – churning out more and more sentence plan trees 23/ 40 Ondřej Dušek . .. . . . . .. . . .. . . .. . . .. . . .. . . . .. . . .. . . .. Natural Language Generation .. . .. . . .. . . • Two main components: • candidate generator : • scorer /ranker for the candidates

  29. • using open_set , close_set – heaps sorted by score . . .. . . .. . . .. . . .. . . .. .. . . . . .. . . .. . Our System Sentence planner Sentence planner – overall – churning out more and more sentence plan trees 23/ 40 Ondřej Dušek . .. .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . . .. Natural Language Generation . .. . . .. . . .. • Two main components: • candidate generator : • scorer /ranker for the candidates • A*-style search • incrementally finding the path • from an empty tree • to a full sentence plan tree which contains all information

  30. . . . .. . . .. . . .. . . .. . . .. . .. .. . . .. . . .. . Our System Sentence planner Sentence planner – overall – churning out more and more sentence plan trees 23/ 40 Ondřej Dušek . .. . .. . . .. . . .. . . .. . . . . . .. . . .. . . . .. . Natural Language Generation .. . . .. . . .. • Two main components: • candidate generator : • scorer /ranker for the candidates • A*-style search • incrementally finding the path • from an empty tree • to a full sentence plan tree which contains all information • using open_set , close_set – heaps sorted by score

  31. • Loop: open_set close_set • viable trees, C + some node(s) • C may be empty open_set 4. check if top score( open_set ) top score( close_set ) • Stop if: a) close_set has better top score than open_set .. . . .. . . . .. Our System . .. . . .. . . Sentence planner Sentence planner – workflow 1. get top-scoring C . put C 2. C candidate generator successors( C ) 3. score C C C put C for d consecutive iterations b) there's nothing lefu on the open list (unlikely) 24/ 40 Ondřej Dušek . .. .. . .. . . .. . . .. . .. . . . .. . . .. . . . .. . . . .. . . .. . . .. . . .. . . .. . . .. . Natural Language Generation • Init: open_set = {empty tree}, close_set = ∅

  32. • viable trees, C + some node(s) • C may be empty open_set 4. check if top score( open_set ) top score( close_set ) • Stop if: a) close_set has better top score than open_set . .. . . .. . . . .. . . .. . . .. . Sentence planner .. . Our System . Sentence planner – workflow 2. C candidate generator successors( C ) 3. score C C C put C for d consecutive iterations b) there's nothing lefu on the open list (unlikely) 24/ 40 Ondřej Dušek .. . .. . . . .. . . .. . .. .. . . .. . . .. . . . Natural Language Generation . . . .. . . .. . . .. . .. . . . .. . . .. • Init: open_set = {empty tree}, close_set = ∅ • Loop: 1. get top-scoring C ← open_set put C → close_set

  33. open_set 4. check if top score( open_set ) top score( close_set ) • Stop if: a) close_set has better top score than open_set . . .. . . .. . . .. . .. . .. .. . . .. . . . . Our System Sentence planner Sentence planner – workflow 3. score C C C put C for d consecutive iterations b) there's nothing lefu on the open list (unlikely) 24/ 40 Ondřej Dušek .. . . .. .. . . .. . . . . . .. . . .. . . . .. .. . . . .. . . .. . . .. . .. . . . .. . Natural Language Generation • Init: open_set = {empty tree}, close_set = ∅ • Loop: 1. get top-scoring C ← open_set put C → close_set 2. C = candidate generator successors( C ) • viable trees, C + some node(s) • C may be empty

  34. 4. check if top score( open_set ) top score( close_set ) • Stop if: a) close_set has better top score than open_set . . .. . . .. . . .. .. . . .. . . .. . . .. . . .. . Our System Sentence planner Sentence planner – workflow for d consecutive iterations b) there's nothing lefu on the open list (unlikely) 24/ 40 Ondřej Dušek . .. . . . . .. . . .. . . .. . . .. . . .. . . .. .. . . .. . . .. . Natural Language Generation . .. . . .. . . • Init: open_set = {empty tree}, close_set = ∅ • Loop: 1. get top-scoring C ← open_set put C → close_set 2. C = candidate generator successors( C ) • viable trees, C + some node(s) • C may be empty 3. score C ′ ∀ C ′ ∈ C put C ′ → open_set

  35. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . Our System Sentence planner Sentence planner – workflow for d consecutive iterations b) there's nothing lefu on the open list (unlikely) 24/ 40 Ondřej Dušek . .. .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . . .. Natural Language Generation . .. . . .. . . .. • Init: open_set = {empty tree}, close_set = ∅ • Loop: 1. get top-scoring C ← open_set put C → close_set 2. C = candidate generator successors( C ) • viable trees, C + some node(s) • C may be empty 3. score C ′ ∀ C ′ ∈ C put C ′ → open_set 4. check if top score( open_set ) > top score( close_set ) • Stop if: a) close_set has better top score than open_set

  36. . . . .. . . .. . . .. . . .. . . .. . .. .. . . .. . . .. . Our System Sentence planner Candidate generator by adding 1 node (at every possible place) 25/ 40 Ondřej Dušek . .. . . . . .. . . .. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. Natural Language Generation • Given a candidate plan tree, generate its successors

  37. . . .. . .. .. . . .. . . .. . . .. . .. . . . .. . . .. . Our System Sentence planner Candidate generator by adding 1 node (at every possible place) 25/ 40 Ondřej Dušek Natural Language Generation . . .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . .. . . .. . .. . . .. . • Given a candidate plan tree, generate its successors t-tree t-tree t-tree t-tree be recommend serve v:fin v:fin v:fin

  38. . . .. . .. .. . . .. . . .. . . .. . .. . . . .. . . .. . Our System Sentence planner Candidate generator by adding 1 node (at every possible place) 25/ 40 Ondřej Dušek Natural Language Generation . . .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . .. . . .. . .. . . .. . • Given a candidate plan tree, generate its successors t-tree t-tree t-tree t-tree be recommend serve v:fin v:fin v:fin

  39. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . Our System Sentence planner Candidate generator by adding 1 node (at every possible place) 25/ 40 Ondřej Dušek Natural Language Generation . .. .. .. . . .. . . .. . . .. . . .. . . . . . . . .. . .. . . .. . . .. . . .. • Given a candidate plan tree, generate its successors t-tree t-tree t-tree t-tree be be be v:fin v:fin v:fin be v:fin restaurant X-name restaurant n:obj n:subj n:subj

  40. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . Our System Sentence planner Candidate generator by adding 1 node (at every possible place) 25/ 40 Ondřej Dušek Natural Language Generation . .. .. .. . . .. . . .. . . .. . . .. . . . . . . . .. . .. . . .. . . .. . . .. • Given a candidate plan tree, generate its successors t-tree t-tree t-tree t-tree be be be v:fin v:fin v:fin be v:fin restaurant X-name restaurant n:obj n:subj n:subj

  41. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . Our System Sentence planner Candidate generator by adding 1 node (at every possible place) 25/ 40 Ondřej Dušek Natural Language Generation . .. .. .. . . .. . . .. . . .. . . .. . . . . . . . .. . . . .. .. . . .. . . .. • Given a candidate plan tree, generate its successors t-tree t-tree t-tree be be be v:fin v:fin v:fin X-name restaurant X-name bar X-name n:subj n:obj n:subj n:obj n:subj

  42. • Limiting by things seen in training data: • + at depth levels • + given input MR • nodes seen with current slot-values • required slot-values for each node .. . . .. . . . . . . .. . . .. . .. . . .. .. Our System Sentence planner Candidate generator – limiting the space 1. t-lemma + formeme combination 2. parent–child combination 3. number of children 4. tree size 5. “weak” compatibility with input MR: 6. “strong” compatibility with input MR: (minimum seen in training data) 26/ 40 Ondřej Dušek . .. . .. . .. . . .. . . . . . .. . . .. . . . .. . . .. . . .. . . .. . .. . . . .. . . .. Natural Language Generation • Number of candidates very high even for small domains • We need to lower the number of “possible” successors

  43. . .. .. . . .. . . . . . .. . . .. . . .. .. 3. number of children Ondřej Dušek 26/ 40 (minimum seen in training data) 6. “strong” compatibility with input MR: 5. “weak” compatibility with input MR: 4. tree size 2. parent–child combination . 1. t-lemma + formeme combination Candidate generator – limiting the space Sentence planner Our System . .. . . .. . . . . .. . . .. . . .. . . .. . . .. . . . .. . . .. . . .. .. . . .. . .. . . Natural Language Generation • Number of candidates very high even for small domains • We need to lower the number of “possible” successors • Limiting by things seen in training data: • + at depth levels • + given input MR • nodes seen with current slot-values • required slot-values for each node

  44. • score • Training: • given m , generate the best tree t top with current weights • update weights if t top t gold (gold-standard) • Update: w feat t gold m feat t top m .. . . .. . . . .. . . . .. . . . .. . . .. .. Our System Sentence planner Scorer Basic perceptron scorer w feat t m w 27/ 40 Ondřej Dušek . .. . .. . .. . . .. . . . .. . .. . . .. . . . . . . .. . . .. . . .. . .. . . . .. . . .. Natural Language Generation • a function: sentence plan tree t , MR m → real-valued score • describes the fitness of t for m

  45. • Training: • given m , generate the best tree t top with current weights • update weights if t top t gold (gold-standard) • Update: w feat t gold m feat t top m . . .. . . .. . . .. .. . . .. . . . . .. . . .. . Our System Sentence planner Scorer Basic perceptron scorer w 27/ 40 Ondřej Dušek .. . .. .. .. . . .. . . . . . .. . . .. . . . . .. . . .. . . .. . . .. . . .. . . .. . Natural Language Generation • a function: sentence plan tree t , MR m → real-valued score • describes the fitness of t for m • score = w ⊤ · feat ( t , m )

  46. • Update: w feat t gold m feat t top m . . .. . . .. . . .. . . .. .. . . . .. . . .. . . .. . Our System Sentence planner Scorer Basic perceptron scorer w 27/ 40 Ondřej Dušek .. . . . . .. .. . . .. . . .. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . Natural Language Generation • a function: sentence plan tree t , MR m → real-valued score • describes the fitness of t for m • score = w ⊤ · feat ( t , m ) • Training: • given m , generate the best tree t top with current weights • update weights if t top ̸ = t gold (gold-standard)

  47. . . . .. . . .. . . .. . . .. . . .. . .. .. . . .. . . .. . Our System Sentence planner Scorer Basic perceptron scorer 27/ 40 Ondřej Dušek . .. . .. . . .. . . .. . . .. . . . . . .. . . .. . . . .. . Natural Language Generation .. . . .. . . .. • a function: sentence plan tree t , MR m → real-valued score • describes the fitness of t for m • score = w ⊤ · feat ( t , m ) • Training: • given m , generate the best tree t top with current weights • update weights if t top ̸ = t gold (gold-standard) • Update: w = w + α · ( feat ( t gold , m ) − feat ( t top , m ))

  48. • + demoting subtrees of wrong generation outputs • Update: find common subtree, start from it and update using gold t i . .. . . .. . . .. . . . . . .. . .. .. . .. . . .. . Our System Sentence planner Differing subtree updates • promoting subtrees of gold-standard trees pairs of subtrees t i top 28/ 40 Ondřej Dušek . .. . . . . .. . . .. . . .. . . .. . . .. . . . .. . . .. . . .. . .. .. . . .. . . Natural Language Generation • Features are global → bigger trees score better • need to promote “promising” incomplete trees

  49. • Update: find common subtree, start from it and update using gold t i . . .. . . .. . . .. . . .. . . . . .. . . .. . . .. . Our System Sentence planner Differing subtree updates pairs of subtrees t i top 28/ 40 Ondřej Dušek .. .. . . . . .. . . .. . . .. . . .. . . .. .. . .. . . .. . . .. Natural Language Generation . . .. . . .. . . • Features are global → bigger trees score better • need to promote “promising” incomplete trees • → promoting subtrees of gold-standard trees • + demoting subtrees of wrong generation outputs

  50. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . Our System Sentence planner Differing subtree updates pairs of subtrees t i top 28/ 40 Ondřej Dušek . .. .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . . . .. .. . . .. . . .. Natural Language Generation • Features are global → bigger trees score better • need to promote “promising” incomplete trees • → promoting subtrees of gold-standard trees • + demoting subtrees of wrong generation outputs • Update: find common subtree, start from it and update using gold , t i

  51. . . .. . . .. . . .. . . .. . . .. .. .. . Sentence planner Natural Language Generation Ondřej Dušek 28/ 40 top pairs of subtrees t i Differing subtree updates Our System . . .. . . .. . . . .. . . . . . .. . . .. . . .. . . . .. . . . .. . . .. .. . .. . . .. . . .. • Features are global → bigger trees score better • need to promote “promising” incomplete trees • → promoting subtrees of gold-standard trees • + demoting subtrees of wrong generation outputs • Update: find common subtree, start from it and update using gold , t i Gold standard ( t gold ): Top generated t top : t-tree t-tree be be v:fin v:fin X restaurant restaurant X n:subj n:obj n:obj n:subj range n:in+X cheap italian adj:attr adj:attr price n:attr moderate adj:attr

  52. . .. .. . . .. . . .. . . .. . . .. . .. . Sentence planner Natural Language Generation Ondřej Dušek 28/ 40 top pairs of subtrees t i Differing subtree updates Our System . . .. . . .. . . . .. . .. . . . .. . . .. . . .. . . .. . . . . .. . . .. .. . .. . . .. . . . • Features are global → bigger trees score better • need to promote “promising” incomplete trees • → promoting subtrees of gold-standard trees • + demoting subtrees of wrong generation outputs • Update: find common subtree, start from it and update using gold , t i Gold standard ( t gold ): Top generated t top : t-tree t-tree Common be subtree be v:fin v:fin ( t c ) X restaurant restaurant X n:subj n:obj n:obj n:subj range n:in+X cheap italian adj:attr adj:attr price n:attr moderate adj:attr

  53. . . .. . . .. . . .. . . .. . . .. . .. . Sentence planner Natural Language Generation Ondřej Dušek 28/ 40 top pairs of subtrees t i Differing subtree updates Our System . . .. . . .. . . .. .. .. . . .. . . .. . . .. . . .. . . . . . .. .. . . . . .. . . .. . . .. . • Features are global → bigger trees score better • need to promote “promising” incomplete trees • → promoting subtrees of gold-standard trees • + demoting subtrees of wrong generation outputs • Update: find common subtree, start from it and update using gold , t i Gold standard ( t gold ): Top generated t top : t-tree t-tree be be v:fin v:fin X restaurant restaurant X n:subj n:obj n:obj n:subj range n:in+X cheap italian adj:attr adj:attr price n:attr t 1 top t 1 gold Differing subtrees for update moderate adj:attr

  54. . . .. . . .. . . .. . . .. . . .. .. .. . Sentence planner Natural Language Generation Ondřej Dušek 28/ 40 top pairs of subtrees t i Differing subtree updates Our System . . .. . . .. . . . .. .. . . .. . . .. . . .. . . .. . . . . . . . . .. . . .. . . .. . . .. .. • Features are global → bigger trees score better • need to promote “promising” incomplete trees • → promoting subtrees of gold-standard trees • + demoting subtrees of wrong generation outputs • Update: find common subtree, start from it and update using gold , t i Gold standard ( t gold ): Top generated t top : t-tree t-tree be be v:fin v:fin X restaurant restaurant X n:subj n:obj n:obj n:subj range n:in+X cheap italian adj:attr adj:attr price n:attr + regular full update moderate adj:attr

  55. • Using expected number of children E c n of a node • Future promise: max 0 E c n • over the whole tree • + multiplied by feature sum • + weighted • used on the open_set , not close_set • not for perceptron updates, not for stopping generation . .. . . .. . . . .. . . .. . . . .. . . .. .. Our System Sentence planner Future promise estimate “how many children are missing to meet the expectation” fc n t c n 29/ 40 Ondřej Dušek . . .. .. . .. . . .. . . . . . .. . . .. . . . .. . . .. . . .. . . .. . . .. . . .. . . .. Natural Language Generation • Further boost for incomplete trees

  56. • Future promise: max 0 E c n • over the whole tree • + multiplied by feature sum • + weighted • used on the open_set , not close_set • not for perceptron updates, not for stopping generation . .. . . .. . . . .. . . .. . . . .. .. . . .. . Our System Sentence planner Future promise estimate “how many children are missing to meet the expectation” fc n t c n 29/ 40 Ondřej Dušek .. . . .. .. . . .. . . . .. . .. . . .. . . . . .. .. . . .. . . .. . . . . . .. . . .. . Natural Language Generation • Further boost for incomplete trees • Using expected number of children E c ( n ) of a node

  57. • over the whole tree • + multiplied by feature sum • + weighted • used on the open_set , not close_set • not for perceptron updates, not for stopping generation . . .. . . .. . . .. . . .. .. . .. .. . . .. . . .. . Our System Sentence planner Future promise estimate “how many children are missing to meet the expectation” 29/ 40 Ondřej Dušek . . . . . . .. . . .. . . .. . . .. . . .. . . .. .. . . .. . . .. . .. . . . .. . . Natural Language Generation • Further boost for incomplete trees • Using expected number of children E c ( n ) of a node • Future promise: ∑ fc = max { 0 , E c ( n ) − c ( n ) } n ∈ t

  58. • used on the open_set , not close_set • not for perceptron updates, not for stopping generation . .. . . .. . . . . . .. .. . .. .. .. . . . . .. . . .. . Our System Sentence planner Future promise estimate “how many children are missing to meet the expectation” 29/ 40 Ondřej Dušek . . .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . .. . Natural Language Generation . .. . . .. . . .. • Further boost for incomplete trees • Using expected number of children E c ( n ) of a node • Future promise: ∑ fc = max { 0 , E c ( n ) − c ( n ) } n ∈ t • over the whole tree • + multiplied by feature sum • + weighted

  59. . . . .. . .. .. . . .. . . .. . . .. . .. .. . . .. . . .. . Our System Sentence planner Future promise estimate “how many children are missing to meet the expectation” 29/ 40 Ondřej Dušek . . . .. . . .. . . .. . . .. . . .. . . . . . .. . . . .. . Natural Language Generation .. . . .. . .. . • Further boost for incomplete trees • Using expected number of children E c ( n ) of a node • Future promise: ∑ fc = max { 0 , E c ( n ) − c ( n ) } n ∈ t • over the whole tree • + multiplied by feature sum • + weighted • used on the open_set , not close_set • not for perceptron updates, not for stopping generation

  60. • Mostly simple, single-purpose, rule-based modules (blocks) • Word inflection: statistical ( Flect ) • Gradual transformation of deep trees into surface dependency • Surface trees are then simply linearized • Works OK: analysis . . .. . . .. . . .. .. . . .. . . .. . . .. . . .. . Our System Surface realizer Surface realizer overview trees synthesis on our data = 89.79% BLEU 30/ 40 Ondřej Dušek . .. . . . . .. . . .. . . .. . . .. . . .. . . . .. . . .. . . .. . .. .. . . .. . . Natural Language Generation • English synthesis pipeline from Treex / TectoMT • domain-independent

  61. • Gradual transformation of deep trees into surface dependency • Surface trees are then simply linearized • Works OK: analysis . . .. . . .. . . .. . . .. . . . . .. . . .. . . .. . Our System Surface realizer Surface realizer overview trees synthesis on our data = 89.79% BLEU 30/ 40 Ondřej Dušek .. .. . . . . .. . . .. . . .. . . .. . . .. .. . .. . . .. . . .. . Natural Language Generation . .. . . .. . . • English synthesis pipeline from Treex / TectoMT • domain-independent • Mostly simple, single-purpose, rule-based modules (blocks) • Word inflection: statistical ( Flect )

  62. • Works OK: analysis . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . Our System Surface realizer Surface realizer overview trees synthesis on our data = 89.79% BLEU 30/ 40 Ondřej Dušek . .. .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . .. . Natural Language Generation . .. . . .. . . .. • English synthesis pipeline from Treex / TectoMT • domain-independent • Mostly simple, single-purpose, rule-based modules (blocks) • Word inflection: statistical ( Flect ) • Gradual transformation of deep trees into surface dependency • Surface trees are then simply linearized

  63. . . . .. . . .. . . .. . . .. . . .. . .. .. . . .. . . .. . Our System Surface realizer Surface realizer overview trees 30/ 40 Ondřej Dušek . .. . .. . . .. . . .. . . .. . . . . . .. . . .. . . . . .. .. . . .. . . .. Natural Language Generation • English synthesis pipeline from Treex / TectoMT • domain-independent • Mostly simple, single-purpose, rule-based modules (blocks) • Word inflection: statistical ( Flect ) • Gradual transformation of deep trees into surface dependency • Surface trees are then simply linearized • Works OK: analysis → synthesis on our data = 89.79% BLEU

  64. • Copy the deep tree (sentence plan) • Determine morphological agreement • Add prepositions and conjunctions • Add articles • Compound verb forms (add auxiliaries) • Punctuation • Word inflection • Capitalization t-tree zone=en_gen jump v:fin cat window n:subj n:through+X . .. . . . . .. . . .. . Surface realization example .. . . .. . Our System Surface realizer . 31/ 40 Ondřej Dušek Natural Language Generation .. . .. . . . .. . . .. . .. .. . . .. . . .. . . .. . . .. . . .. . . .. . . . . .. . . .. . . .. . • Realizer steps (simplified):

  65. • Determine morphological agreement • Add prepositions and conjunctions • Add articles • Compound verb forms (add auxiliaries) • Punctuation • Word inflection • Capitalization . . .. . . .. . . .. .. . .. .. . . . . .. . . .. . Our System Surface realizer Surface realization example 31/ 40 Ondřej Dušek Natural Language Generation . . .. .. .. . . .. . . . . . .. . . .. . . . . .. .. . .. . . .. . . . . . . .. . . .. • Realizer steps (simplified): • Copy the deep tree (sentence plan) t-tree zone=en_gen jump v:fin cat window n:subj n:through+X

  66. • Add prepositions and conjunctions • Add articles • Compound verb forms (add auxiliaries) • Punctuation • Word inflection • Capitalization . .. . . .. . . .. . . .. .. . . .. .. . . .. . . .. . Our System Surface realizer Surface realization example 31/ 40 Ondřej Dušek Natural Language Generation . . . . . . . . . .. . . .. . . .. . . .. .. . . .. . . .. . .. .. . . . .. . . .. . • Realizer steps (simplified): • Copy the deep tree (sentence plan) • Determine morphological agreement t-tree zone=en_gen jump v:fin cat window n:subj n:through+X

  67. • Add articles • Compound verb forms (add auxiliaries) • Punctuation • Word inflection • Capitalization . .. . . .. . . .. . . .. .. . . .. .. . . .. . . .. . Our System Surface realizer Surface realization example 31/ 40 Ondřej Dušek Natural Language Generation . . . . . . . . . .. . . .. . . .. . . .. .. . . .. . . .. . . .. .. . .. . . .. . . • Realizer steps (simplified): • Copy the deep tree (sentence plan) • Determine morphological agreement • Add prepositions and conjunctions t-tree zone=en_gen jump v:fin cat window n:subj n:through+X

  68. • Compound verb forms (add auxiliaries) • Punctuation • Word inflection • Capitalization . . .. . . .. . .. .. . . .. . . . . .. . . .. . . .. . Our System Surface realizer Surface realization example 31/ 40 Ondřej Dušek Natural Language Generation .. . . . . . .. .. . .. . . .. . . .. . . .. . . .. . . .. . . .. . . . . .. . . .. • Realizer steps (simplified): • Copy the deep tree (sentence plan) • Determine morphological agreement • Add prepositions and conjunctions • Add articles t-tree zone=en_gen jump v:fin cat window n:subj n:through+X

  69. • Punctuation • Word inflection • Capitalization . . .. . . .. . . .. . .. .. . . . . .. . . .. . . .. . Our System Surface realizer Surface realization example 31/ 40 Ondřej Dušek Natural Language Generation .. . . . . . .. . .. .. . . .. . . .. . . .. . . . .. . . .. .. . . . .. . . .. . . • Realizer steps (simplified): • Copy the deep tree (sentence plan) • Determine morphological agreement • Add prepositions and conjunctions • Add articles • Compound verb forms (add auxiliaries) t-tree zone=en_gen jump v:fin cat window n:subj n:through+X

  70. • Word inflection • Capitalization . .. . . .. . . .. . . .. . . .. .. .. . . . . .. . . .. . Our System Surface realizer Surface realization example 31/ 40 Ondřej Dušek Natural Language Generation . . .. .. . . .. . . .. . . . . . .. . . .. . . .. .. . . . . .. . . .. . . .. . • Realizer steps (simplified): • Copy the deep tree (sentence plan) • Determine morphological agreement • Add prepositions and conjunctions • Add articles • Compound verb forms (add auxiliaries) • Punctuation t-tree zone=en_gen jump v:fin cat window n:subj n:through+X

  71. • Capitalization . . .. . . .. .. . .. . . .. . . .. .. . . . . .. . . .. . Our System Surface realizer Surface realization example 31/ 40 Ondřej Dušek Natural Language Generation . . .. .. . . .. . . .. . . . . . .. . . .. . . .. . . .. . . .. . . .. . . .. . • Realizer steps (simplified): • Copy the deep tree (sentence plan) • Determine morphological agreement • Add prepositions and conjunctions • Add articles • Compound verb forms (add auxiliaries) • Punctuation • Word inflection t-tree zone=en_gen jump v:fin cat window n:subj n:through+X

  72. . . . .. . . .. .. . .. . . .. . . .. . .. .. . . .. . . .. . Our System Surface realizer Surface realization example 31/ 40 Ondřej Dušek Natural Language Generation . . . .. . . .. . . .. . . .. . . .. . . . . .. .. . .. . . . . . .. . . .. . • Realizer steps (simplified): • Copy the deep tree (sentence plan) • Determine morphological agreement • Add prepositions and conjunctions • Add articles • Compound verb forms (add auxiliaries) • Punctuation • Word inflection • Capitalization t-tree zone=en_gen jump v:fin cat window n:subj n:through+X

  73. • Alignment provided, but we don't use it • “Non-enumerable” information replaced by “X” symbol • restaurant names, postcodes, phone numbers etc. . .. . . .. . . . . . .. . . .. .. .. . . . . .. . . .. . Our System Experiments Experiments – data set 32/ 40 Ondřej Dušek .. . .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . .. . . .. . . .. . . .. Natural Language Generation • Restaurant recommendations from the BAGEL generator • restaurant location, food type, etc. • 404 utterances for 202 input dialogue acts (DAs) • two paraphrases for each DA

  74. • “Non-enumerable” information replaced by “X” symbol • restaurant names, postcodes, phone numbers etc. . .. . . .. . . . . . .. . . .. .. .. . . . . .. . . .. . Our System Experiments Experiments – data set 32/ 40 Ondřej Dušek .. . .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . . . .. .. . . .. . . .. Natural Language Generation • Restaurant recommendations from the BAGEL generator • restaurant location, food type, etc. • 404 utterances for 202 input dialogue acts (DAs) • two paraphrases for each DA • Alignment provided, but we don't use it

  75. . .. .. .. . . .. . . .. . . .. . . . .. . .. . . .. . . .. . Our System Experiments Experiments – data set 32/ 40 Ondřej Dušek . . . .. . . .. . . .. . . .. . . . . . .. . . .. . . . . .. .. . . .. . . .. Natural Language Generation • Restaurant recommendations from the BAGEL generator • restaurant location, food type, etc. • 404 utterances for 202 input dialogue acts (DAs) • two paraphrases for each DA • Alignment provided, but we don't use it • “Non-enumerable” information replaced by “X” symbol • restaurant names, postcodes, phone numbers etc.

  76. • Basic feature types: • tree properties (size, depth…) • tree + input DA (nodes per slot-value pair…) • node features • input DA feautres (slots, values, pairs of slots) • node + input DA features • repeat features (repeated nodes/slots/values) • dependency features (parent-child) • siblings features (+DA) • bigram features (+DA) • Typical case: counts over whole tree • normalized .. . .. . . . . . .. .. . . . . . .. . . .. .. Our System Experiments Experiments – features 33/ 40 Ondřej Dušek . . .. .. . .. . . .. . . . . . .. . . .. . . . .. . . .. . . .. . . .. . . .. . . .. . . .. Natural Language Generation • Tailored for the input MR format

  77. • Typical case: counts over whole tree • normalized . .. . . .. . . . . . .. .. . .. .. .. . . . . .. . . .. . Our System Experiments Experiments – features 33/ 40 Ondřej Dušek . . .. .. . . .. . . .. . . .. . . .. . . . . . . .. . . . .. Natural Language Generation . .. . . .. . . .. • Tailored for the input MR format • Basic feature types: • tree properties (size, depth…) • tree + input DA (nodes per slot-value pair…) • node features • input DA feautres (slots, values, pairs of slots) • node + input DA features • repeat features (repeated nodes/slots/values) • dependency features (parent-child) • siblings features (+DA) • bigram features (+DA)

  78. . .. . .. . .. .. . . .. . . .. . . . .. . .. . . .. . . .. . Our System Experiments Experiments – features 33/ 40 Ondřej Dušek . . . .. . . .. . . .. . . . . . .. . . .. . . .. . . . .. . Natural Language Generation .. . . .. . .. . • Tailored for the input MR format • Basic feature types: • tree properties (size, depth…) • tree + input DA (nodes per slot-value pair…) • node features • input DA feautres (slots, values, pairs of slots) • node + input DA features • repeat features (repeated nodes/slots/values) • dependency features (parent-child) • siblings features (+DA) • bigram features (+DA) • Typical case: counts over whole tree • normalized

  79. • less than BAGEL 's ~ 67% BLEU • But: • we do not use alignments • our generator has to know when to stop (whether all . .. .. . . .. . . .. . . Our System . . .. . . . Setup Results 4.876 Ondřej Dušek 34/ 40 information is already included) 5.231 59.89 + future promise 58.70 Results + diff-tree updates 4.643 54.24 basic perceptron NIST BLEU . .. .. . . .. . . .. . . .. . .. .. . . .. . . .. . . . Natural Language Generation . . . . .. . . .. . .. . . . .. . . .. . .. • Using 10-fold cross-validation, measuring BLEU/NIST • training DAs never used for testing • using 2 paraphrases for BLEU/NIST measurements

  80. • less than BAGEL 's ~ 67% BLEU • But: • we do not use alignments • our generator has to know when to stop (whether all . .. .. . . .. . . .. . . Our System . . .. . . . Setup Results 4.876 Ondřej Dušek 34/ 40 information is already included) 5.231 59.89 + future promise 58.70 Results + diff-tree updates 4.643 54.24 basic perceptron NIST BLEU . .. .. . . .. . . .. . . .. . .. .. . . .. . . .. . . . Natural Language Generation . . . . .. . . .. . .. . . . .. . . .. . .. • Using 10-fold cross-validation, measuring BLEU/NIST • training DAs never used for testing • using 2 paraphrases for BLEU/NIST measurements

  81. • But: • we do not use alignments • our generator has to know when to stop (whether all . .. .. . . .. . . .. . . Our System . . .. . . . Results Results 4.876 Ondřej Dušek 34/ 40 information is already included) 5.231 59.89 + future promise 58.70 . + diff-tree updates 4.643 54.24 basic perceptron NIST BLEU Setup .. .. . . . . .. . . .. . .. .. . . .. . . .. . . .. Natural Language Generation . . . . .. . . .. . . .. . . .. . . .. . . .. • Using 10-fold cross-validation, measuring BLEU/NIST • training DAs never used for testing • using 2 paraphrases for BLEU/NIST measurements • less than BAGEL 's ~ 67% BLEU

  82. . . . . .. . . .. . .. . . . .. . . .. . .. Our System .. 58.70 Ondřej Dušek 34/ 40 information is already included) 5.231 59.89 + future promise 4.876 + diff-tree updates Results 4.643 54.24 basic perceptron NIST BLEU Setup Results . .. . .. . .. . . .. . . . . . .. . . .. . . . Natural Language Generation .. . .. . . .. . . .. . . .. . . .. . . .. . • Using 10-fold cross-validation, measuring BLEU/NIST • training DAs never used for testing • using 2 paraphrases for BLEU/NIST measurements • less than BAGEL 's ~ 67% BLEU • But: • we do not use alignments • our generator has to know when to stop (whether all

  83. . Our System Generated X is a French and continental restaurant near X. Reference food=Continental, food=French) inform(name=X-name, type=placetoeat, eattype=restaurant, near=X-near, Input DA Example Outputs Results . Input DA .. . . .. . . .. . . X is a French and continental restaurant near X. inform(name=X-name, type=placetoeat, area=riverside, near=X-near, . X is a restaurant in the X area. Ondřej Dušek 35/ 40 X is a French restaurant in the riverside area which serves French food. Generated X is a French restaurant on the riverside. Reference area=riverside, food=French) inform(name=X-name, type=placetoeat, eattype=restaurant, Input DA Generated eattype=restaurant) X is a moderately priced restaurant in X. Reference pricerange=moderate, eattype=restaurant) inform(name=X-name, type=placetoeat, area=X-area, Input DA X is a restaurant in the riverside area near X. Generated X restaurant is near X on the riverside. Reference .. . .. . . . .. . . .. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . . .. . .. .. . . .. . . .. . . Natural Language Generation

  84. . Our System Generated X is a French and continental restaurant near X. Reference food=Continental, food=French) inform(name=X-name, type=placetoeat, eattype=restaurant, near=X-near, Input DA Example Outputs Results . Input DA .. . . .. . . .. . . X is a French and continental restaurant near X. inform(name=X-name, type=placetoeat, area=riverside, near=X-near, . X is a restaurant in the X area. Ondřej Dušek 35/ 40 X is a French restaurant in the riverside area which serves French food. Generated X is a French restaurant on the riverside. Reference area=riverside, food=French) inform(name=X-name, type=placetoeat, eattype=restaurant, Input DA Generated eattype=restaurant) X is a moderately priced restaurant in X. Reference pricerange=moderate, eattype=restaurant) inform(name=X-name, type=placetoeat, area=X-area, Input DA X is a restaurant in the riverside area near X. Generated X restaurant is near X on the riverside. Reference .. . .. . . . .. . . .. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . . .. . .. .. . . .. . . .. . . Natural Language Generation

  85. . Our System Generated X is a French and continental restaurant near X. Reference food=Continental, food=French) inform(name=X-name, type=placetoeat, eattype=restaurant, near=X-near, Input DA Example Outputs Results . Input DA .. . . .. . . .. . . X is a French and continental restaurant near X. inform(name=X-name, type=placetoeat, area=riverside, near=X-near, . X is a restaurant in the X area. Ondřej Dušek 35/ 40 X is a French restaurant in the riverside area which serves French food. Generated X is a French restaurant on the riverside. Reference area=riverside, food=French) inform(name=X-name, type=placetoeat, eattype=restaurant, Input DA Generated eattype=restaurant) X is a moderately priced restaurant in X. Reference pricerange=moderate, eattype=restaurant) inform(name=X-name, type=placetoeat, area=X-area, Input DA X is a restaurant in the riverside area near X. Generated X restaurant is near X on the riverside. Reference .. . .. . . . .. . . .. . . .. . . .. . . .. . . .. . . .. . .. . . . .. . . .. . . .. . .. .. . . .. . . .. . . Natural Language Generation

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend