extractive summarization with swap net sentences and
play

Extractive Summarization with SWAP-NET: Sentences and Words from - PowerPoint PPT Presentation

Extractive Summarization with SWAP-NET: Sentences and Words from Alternating Pointer Networks Aishwarya Jadhav Vaibhav Rajan Indian Institute of Science School of Computing Bangalore, India National University of Singapore Extractive


  1. Extractive Summarization with SWAP-NET: Sentences and Words from Alternating Pointer Networks Aishwarya Jadhav Vaibhav Rajan Indian Institute of Science School of Computing Bangalore, India National University of Singapore

  2. Extractive Summarization Select salient sentences from input document to create a summary S 1 S i 1 S 2 • Supervised extractive summarization for single document inputs S i m S n OUTPUT Summary 1 ≤ i k ≤ n INPUT Document with sentences S 1 , S 2 ,.., S n

  3. Our Contribution A Deep Learning Architecture for training an extractive summarizer: SWAP-NET S 1 • Unlike previous methods, SWAP-NET uses keywords for sentence selection S i 1 • Predicts both important words and S 2 sentences in document • Two-level Encoder-Decoder Attention model S i m • Outperform state of the art extractive S n summarisers. OUTPUT Summary 1 ≤ i k ≤ n INPUT Document with sentences S 1 , S 2 ,.., S n

  4. Extractive Summarization Methods Recent extractive summarization methods

  5. Extractive Summarization Methods Recent extractive summarization methods • NN (Cheng and Lapata, 2016) Sentence Label Pre-trained word Sentence Encoding Sentence encodings Prediction embeddings wrt words in it wrt other sentences (with decoder) Jianpeng Cheng and Mirella Lapata. 2016. Neural summarization by extracting sentences and words. 54th Annual Meeting of the Association for Computational Linguistics.

  6. Extractive Summarization Methods Recent extractive summarization methods • NN (Cheng and Lapata, 2016) Sentence Label Pre-trained word Sentence Encoding Sentence encodings Prediction embeddings wrt words in it wrt other sentences (with decoder) Sentence Label • SummaRuNNer (Nallapati et al., 2017) Prediction Pre-trained word Word Encodings Sentence Encoding Sentence Encodings Document Encoding embeddings wrt other words wrt words in it wrt other sentences wrt its sentences Ramesh Nallapati, Feifei Zhai, and Bowen Zhou. 2017. Summarunner: A recurrent neural network based sequence model for extractive summarization of docments. In Association for the Advancement of Artificial Intelligence, pages 3075–3081. Jianpeng Cheng and Mirella Lapata. 2016. Neural summarization by extracting sentences and words. 54th Annual Meeting of the Association for Computational Linguistics.

  7. Extractive Summarization Methods Recent extractive summarization methods • NN (Cheng and Lapata, 2016) Sentence Label Pre-trained word Sentence Encoding Sentence encodings Prediction embeddings wrt words in it wrt other sentences (with decoder) Sentence Label • SummaRuNNer (Nallapati et al., 2017) Prediction Pre-trained word Word Encodings Sentence Encoding Sentence Encodings Document Encoding embeddings wrt other words wrt words in it wrt other sentences wrt its sentences • Both assume saliency of sentence s depends on salient sentences appearing before s Ramesh Nallapati, Feifei Zhai, and Bowen Zhou. 2017. Summarunner: A recurrent neural network based sequence model for extractive summarization of docments. In Association for the Advancement of Artificial Intelligence, pages 3075–3081. Jianpeng Cheng and Mirella Lapata. 2016. Neural summarization by extracting sentences and words. 54th Annual Meeting of the Association for Computational Linguistics.

  8. Intuition Behind Approach Question : Which sentence should be considered salient (part of summary)? • Our hypothesis: saliency of a sentence depends on both salient sentences and words appearing before that sentence in the document • Similar to graph based models by Wan et al. (2007) • Along with labelling sentences we also label words to determine their saliency • Moreover, saliency of a word depends on previous salient words and sentences Xiaojun Wan, Jianwu Yang, and Jianguo Xiao. 2007. Towards an iterative reinforcement approach for simultaneous document summarization and keyword extraction. In Proceedings of the 45th annual meeting of the association of computational linguistics , pages 552–559.

  9. Intuition Behind Approach Three types of Interactions: • Sentence-Sentence Interaction • Word-Word Interaction • Sentence-Word Interaction

  10. Intuition: Interaction Between Sentences A sentence should be salient if it is heavily linked with other salient sentences Sentence - Sentence S1 S2 S3 V6 V5 V4 V3 V2 V1

  11. Intuition: Interaction Between Words A word should be salient if it is heavily linked with other salient words S1 S2 S3 V6 V5 V4 V3 V2 V1 Word-Word

  12. Intuition: Words and Sentences Interaction A sentence should be salient if it contains many salient words A word should be salient if it appears in many salient sentences Sentence-Word S1 S2 S3 V6 V5 V4 V3 V2 V1

  13. Intuition: Words and Sentences Interaction Generate extractive summary using both important words and sentences Sentence - Sentence Sentence-Word S1 S2 S3 V6 V5 V4 V3 V2 V1 Important Sentences: S3 Word-Word Important Words: V2, V3

  14. Keyword Extraction and Sentence Extraction • Sentence to Sentence Interaction as Sentence Extraction • Word to Word Interaction as Word Extraction • For discrete sequences, pointer networks have been successfully used to learn how to select positions from an input sequence • We use two pointer networks one at word-level and another at sentence-level

  15. Pointer Network Pointer network (Vinyals et al., 2015), • Encoder-Decoder architecture with Attention • Attention mechanism is used to select one of the inputs at each decoding step • Thus, e ff ectively pointing to an input Attention Vector 3 2 e2 e1 e3 e4 d1 d2 Input Output Indices (R): 2,3 x1 x2 x3 x4 (X): Decoder Encoder Oriol Vinyals, Meire Fortunato, and Navdeep Jaitly. 2015. Pointer networks. In Advances in Neural Information Processing Systems , pages 2692–2700.

  16. Three Interactions Sentence-Level Pointer Network ? Sentence - Sentence Sentence-Word S1 S2 S3 V6 V5 V4 V3 V2 V1 Word-Word Word-Level Pointer Network

  17. Three Interactions: SWAP-NET Sentence-Level Pointer Network Sentence - Sentence Sentence-Word A Mechanism to Combine Word Level Attentions and Generate Summary Sentence Level Attentions Word-Word Word-Level Pointer Network

  18. Questions Q1 : How can the two attentions be combined? Q2 : How can the summaries be generated considering both the attentions? Q1 Q2 A Mechanism to Combine ? ? Sentence-Word Word Level Attentions and Generate Summary Sentence Level Attentions

  19. Three Interactions: SWAP-NET Sentence-Level Pointer Network ? Sentence - Sentence Sentence-Word S1 S2 S3 V6 V5 V4 V3 V2 V1 Word-Word Word-Level Pointer Network

  20. SWAP-NET Architecture: Word-Level Pointer Network Similar to Pointer Network, • The word encoder is bi-directional LSTM • Word-level decoder learns to point to important words Word Word Decoder Encoder E E E E E D D D W W W W W W W W 1 2 3 4 5 1 2 3 w1 w2 w3 w4 w5

  21. SWAP-NET Architecture: Word-Level Pointer Network Probability of word i, • Purple line: attention vector given as input at decoding step j to each decoding step • Sum of word encodings weighted by attention probabilities generated in previous step Word Attention w1 w2 w3 w4 w5 Word Attention E E E E E D D D W W W W W W W W Vector 1 2 3 4 5 1 2 3 w1 w2 w3 w4 w5

  22. Three Interactions: SWAP-NET Sentence-Level Pointer Network ? Sentence - Sentence Sentence-Word S1 S2 S3 V6 V5 V4 V3 V2 V1 Word-Word Word-Level Pointer Network

  23. SWAP-NET Architecture: Sentence-Level Hierarchical Pointer Network Sentence is represented by encoding of last word of that sentence Sentence Sentence E E D D D Encoder Decoder S S S S S 1 2 1 2 3 s1 s2 Word Word E E E E E D D D Decoder Encoder W W W W W W W W 1 2 3 4 5 1 2 3 w1 w2 w3 w4 w5

  24. SWAP-NET Architecture: Sentence-Level Hierarchical Pointer Network Attention vectors are sum of sentence encodings weighted by attention probabilities by previous decoding step Sentence Attention E E D D D Vector S S S S S 1 2 1 2 3 s1 s2 Probability of sentence k, at decoding step j Sentence Attention E E E E E D D D W W W W W W W W 1 2 3 4 5 1 2 3 w1 w2 w3 w4 w5

  25. Combining Sentence Attention and Word Attention Q1 : How can the two attentions be combined? Sentences S1 S2 S3 V1 V2 V2 V5 V4 V6 V4 V3 V2 Words A document with three sentences and corresponding words is shown

  26. Sentence and Word Interactions Possible Solution: Step 1: Hold sentence processing. Then group all words and determine their saliency sequentially S1 S2 S3 V1 V2 V2 V5 V4 V6 V4 V3 V2

  27. Sentence and Word Interactions Possible Solution: Step 2: Using output of step 1, i.e., using keywords, process sentences to determine salient sentences S1 S2 S3 V1 V2 V2 V5 V4 V6 V4 V3 V2 INCOMPLETE SOLUTION : This methods processes sentence depending on words but does not use sentences for processing words.

  28. Sentence and Word Interactions Solution : Group each sentence and its words separately and process them sequentially S1 S2 S3 V1 V2 V2 V5 V4 V6 V4 V3 V2

  29. Sentence and Word Interactions Step1: Hold sentence processing. Determine saliency of words in S1 S1 S2 S3 V1 V2 V2 V5 V4 V6 V4 V3 V2

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend