semantics as a foreign language
play

Semantics as a Foreign Language Gabriel Stanovsky and Ido Dagan - PowerPoint PPT Presentation

Semantics as a Foreign Language Gabriel Stanovsky and Ido Dagan EMNLP 2018 Semantic Dependency Parsing (SDP) A collection of three semantic formalisms (Oepen et al., 2014;2015) Semantic Dependency Parsing (SDP) A collection of three


  1. Semantics as a Foreign Language Gabriel Stanovsky and Ido Dagan EMNLP 2018

  2. Semantic Dependency Parsing (SDP) ● A collection of three semantic formalisms (Oepen et al., 2014;2015)

  3. Semantic Dependency Parsing (SDP) ● A collection of three semantic formalisms (Oepen et al., 2014;2015) a. DM (derived from MRS) (Copestake et al., 1999, Flickinger, 2000)

  4. Semantic Dependency Parsing (SDP) ● A collection of three semantic formalisms (Oepen et al., 2014;2015) a. DM (derived from MRS) b. Prague Semantic Dependencies (PSD) (Hajic et al., 2012)

  5. Semantic Dependency Parsing (SDP) ● A collection of three semantic formalisms (Oepen et al., 2014;2015) a. DM (derived from MRS) b. Prague Semantic Dependencies (PSD) c. Predicate Argument Structures (PAS) (Miyao et al., 2014)

  6. Semantic Dependency Parsing (SDP) ● A collection of three semantic formalisms (Oepen et al., 2014;2015) a. DM (derived from MRS) b. Prague Semantic Dependencies (PSD) c. Predicate Argument Structures (PAS) ● Aim to capture semantic predicate-argument relations

  7. Semantic Dependency Parsing (SDP) ● A collection of three semantic formalisms (Oepen et al., 2014;2015) a. DM (derived from MRS) b. Prague Semantic Dependencies (PSD) c. Predicate Argument Structures (PAS) ● Aim to capture semantic predicate-argument relations ● Represented in a graph structure

  8. Semantic Dependency Parsing (SDP) ● A collection of three semantic formalisms (Oepen et al., 2014;2015) a. DM (derived from MRS) b. Prague Semantic Dependencies (PSD) c. Predicate Argument Structures (PAS) ● Aim to capture semantic predicate-argument relations ● Represented in a graph structure a. Nodes: single words from the sentence

  9. Semantic Dependency Parsing (SDP) ● A collection of three semantic formalisms (Oepen et al., 2014;2015) a. DM (derived from MRS) b. Prague Semantic Dependencies (PSD) c. Predicate Argument Structures (PAS) ● Aim to capture semantic predicate-argument relations ● Represented in a graph structure a. Nodes: single words from the sentence b. Labeled edges: semantic relations, according to the paradigm

  10. Outline

  11. Outline ● SDP as Machine Translation ○ Different formalisms as foreign languages ○ Motivation : downstream tasks, inter-task analysis, extendable framework

  12. Outline ● SDP as Machine Translation ○ Different formalisms as foreign languages ○ Motivation : downstream tasks, inter-task analysis, extendable framework ○ Previous work explored the relation between MT and semantics (Wong and Mooney, 2007), (Vinyals et al., 2015), (Flanigan et al., 2016)

  13. Outline ● SDP as Machine Translation ○ Different formalisms as foreign languages ○ Motivation : downstream tasks, inter-task analysis, extendable framework ○ Previous work explored the relation between MT and semantics (Wong and Mooney, 2007), (Vinyals et al., 2015), (Flanigan et al., 2016) ● Model ○ Seq2Seq ○ Directed graph linearization

  14. Outline ● SDP as Machine Translation ○ Different formalisms as foreign languages ○ Motivation : downstream tasks, inter-task analysis, extendable framework ○ Previous work explored the relation between MT and semantics (Wong and Mooney, 2007), (Vinyals et al., 2015), (Flanigan et al., 2016) ● Model ○ Seq2Seq ○ Directed graph linearization ● Results ○ Raw text to SDP (near state-of-the-art) ○ Novel inter-task analysis

  15. Outline ● SDP as Machine Translation ○ Different formalisms as foreign languages ○ Motivation : downstream tasks, inter-task analysis, extendable framework ○ Previous work explored the relation between MT and semantics (Wong and Mooney, 2007), (Vinyals et al., 2015), (Flanigan et al., 2016) ● Model ○ Seq2Seq ○ Linearization ● Results ○ Raw text -> SDP (near state-of-the-art) ○ Novel inter-task analysis

  16. Semantic Dependencies as MT Source Raw sentence Target

  17. Semantic Dependencies as MT Source Raw sentence Grammar as a foreign language Syntax Target

  18. Semantic Dependencies as MT Source Raw sentence Grammar as a foreign language T h i s w o r k Syntax SDP Target

  19. Semantic Dependencies as MT ● Standard MTL: 3 tasks Raw sentence PSD PAS DM

  20. Semantic Dependencies as MT ● Standard MTL: 3 tasks Raw sentence PSD PAS DM ● Inter-task translation (9 tasks)

  21. Outline ● SDP as Machine Translation ○ Motivation: downstream tasks ○ Different formalisms as foreign languages ● Model ○ Seq2Seq ○ Linearization ● Results ○ Raw text -> SDP (near state-of-the-art) ○ Novel inter-task analysis

  22. Our Model Ⅰ : Raw -> SDP x ● Seq2Seq translation model: ○ Bi-LSTM encoder-decoder with attention Linear DM the cat sat on the mat <from: RAW> <to: DM>

  23. Our Model Ⅰ : Raw -> SDP x ● Seq2Seq translation model: ○ Bi-LSTM encoder-decoder with attention Linear DM the cat sat on the mat <from: RAW> <to: DM> Special from and to symbols

  24. Our Model Ⅱ : SDP y -> SDP x ● Seq2Seq translation model: ○ Bi-LSTM encoder-decoder with attention Linear DM <from: PSD> <to: DM> Linear PSD Special from and to symbols

  25. Our Model Seq2seq prediction requires a 1:1 linearization function Linear SDP x Linear SDP y <from: SDP y > <to: SDP x >

  26. Linearization: Background ● Previous work used bracketed tree linearization (ROOT (NP (NNP Bob )NNP )NP (VP messaged (NP Alice )NP )VP )ROOT (Vinyals et al., 2015; Konstas et al., 2017; Buys and Blunsom, 2017)

  27. Linearization: Background ● Previous work used bracketed tree linearization (ROOT (NP (NNP Bob )NNP )NP (VP messaged (NP Alice )NP )VP )ROOT (Vinyals et al., 2015; Konstas et al., 2017; Buys and Blunsom, 2017)

  28. Linearization: Background ● Previous work used bracketed tree linearization (ROOT (NP (NNP Bob )NNP )NP (VP messaged (NP Alice )NP )VP )ROOT (Vinyals et al., 2015; Konstas et al., 2017; Buys and Blunsom, 2017)

  29. Linearization: Background ● Previous work used bracketed tree linearization (ROOT (NP (NNP Bob )NNP )NP (VP messaged (NP Alice )NP )VP )ROOT (Vinyals et al., 2015; Konstas et al., 2017; Buys and Blunsom, 2017)

  30. Linearization: Background ● Previous work used bracketed tree linearization (ROOT (NP (NNP John )NNP )NP (VP messaged (NP Alice )NP )VP )ROOT (Vinyals et al., 2015; Konstas et al., 2017; Buys and Blunsom, 2017)

  31. Linearization: Background ● Previous work used bracketed tree linearization (ROOT (NP (NNP John )NNP )NP (VP messaged (NP Alice )NP )VP )ROOT (Vinyals et al., 2015; Konstas et al., 2017; Buys and Blunsom, 2017)

  32. Linearization: Background ● Previous work used bracketed tree linearization (ROOT (NP (NNP John )NNP )NP (VP messaged (NP Alice )NP )VP )ROOT (Vinyals et al., 2015; Konstas et al., 2017; Buys and Blunsom, 2017) ● Depth-first representation doesn’t directly apply to SDP graphs ○ Non-connected components ○ Re-entrencies

  33. SDP Linearization (Connectivity) ● Problem: No single root from which to start linearization

  34. SDP Linearization (Connectivity) ● Problem: No single root from which to start linearization

  35. SDP Linearization (Connectivity) ● Problem: No single root from which to start linearization ● Solution : Artificial SHIFT edges between non-connected adjacent words ○ All nodes are now reachable from the first word

  36. SDP Linearization (Re-entrancies) ● Re-entrancies require a 1:1 node representation

  37. SDP Linearization (Re-entrencies) ● Re-entrancies require a 1:1 node representation

  38. SDP Linearization (Re-entrencies) ● Re-entrancies require a 1:1 node representation (relative index / surface form)

  39. SDP Linearization (Re-entrencies) ● Re-entrancies require a 1:1 node representation (relative index / surface form) 0/couch-potato

  40. SDP Linearization (Re-entrencies) ● Re-entrancies require a 1:1 node representation (relative index / surface form) 0/couch-potato compound +1/jocks

  41. SDP Linearization (Re-entrencies) ● Re-entrancies require a 1:1 node representation (relative index / surface form) 0/couch-potato compound +1/jocks shift +1/watching

  42. SDP Linearization (Re-entrencies) ● Re-entrancies require a 1:1 node representation (relative index / surface form) 0/couch-potato compound +1/jocks shift +1/watching ARG1 -1/jocks

  43. Outline ● SDP as Machine Translation ○ Motivation: downstream tasks ○ Different formalisms as foreign languages ● Model ○ Linearization ○ Dual Encoder-Single decoder Seq2Seq ● Results ○ Raw text -> SDP (near state-of-the-art) ○ Novel inter-task analysis

  44. Experimental Setup ● Train samples per task: 35,657 sentences (Oepen et al., 2015) ○ 9 translation tasks

  45. Experimental Setup ● Train samples per task: 35,657 sentences (Oepen et al., 2015) ○ 9 translation tasks ● Total training samples: 320,913 source-target pairs

  46. Experimental Setup ● Train samples per task: 35,657 sentences (Oepen et al., 2015) ○ 9 translation tasks ● Total training samples: 320,913 source-target pairs ● Trained in batches between the 9 different tasks

  47. Evaluations:RAW → SDP (x) Labeled F1 score

  48. Evaluations:RAW → SDP (x) Labeled F1 score

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend