Outline Problem Description MASC Architecture MASC Results - - PowerPoint PPT Presentation

outline
SMART_READER_LITE
LIVE PREVIEW

Outline Problem Description MASC Architecture MASC Results - - PowerPoint PPT Presentation

Multiple Alternative Sentence Compressions (MASC) A Framework for Automatic Summarization Nitin Madnani, David Zajic, Bonnie Dorr Necip Fazil Ayan, Jimmy Lin University of Maryland, College Park 1 Outline Problem Description MASC


slide-1
SLIDE 1

1

Multiple Alternative Sentence Compressions (MASC)

A Framework for Automatic Summarization

Nitin Madnani, David Zajic, Bonnie Dorr

Necip Fazil Ayan, Jimmy Lin University of Maryland, College Park

slide-2
SLIDE 2

2

Outline

  • Problem Description
  • MASC Architecture
  • MASC Results
  • Improving Candidate Selection
  • Summary & Future Work
slide-3
SLIDE 3

3

Problem Description

  • Sentence-level extractive summarization

– Source sentences contain mixture of relevant/non- relevant, novel/redundant information.

  • Compression

– Single output compression can’t provide best compression of each sentence for every user need.

  • Multiple Alternative Sentence Compression

– Generation of multiple candidate compressions of source sentences. – Feature-based selection to choose among candidates.

slide-4
SLIDE 4

4

Outline

  • Problem Description
  • MASC Architecture
  • MASC Results
  • Improving Candidate Selection
  • Summary & Future Work
slide-5
SLIDE 5

5

MASC Architecture

Sentence Filtering Sentence Compression Candidate Selection Sentences Candidates Task-Specific Features (e.g. query) Documents Summary

HMM Hedge Trimmer Topiary

(Zajic et al., 2005) (Zajic et al., 2006)

slide-6
SLIDE 6

6

HMM Hedge Architecture

Part of Speech Tagger1 HMM Hedge Sentence Sentence with Verb Tags

             

VERB VERB

Compressions

                    

Headline Language Model Story Language Model

Language models based on 242,918 AP headlines and stories from Tipster Corpus

1TreeTagger (Schmid, 1994)

slide-7
SLIDE 7

7

HMM Hedge Multiple Alternative Compressions

  • Calculate best compression at each word-length

from 5 to 15 words

  • Calculate 5 best compressions at each word

length

slide-8
SLIDE 8

8

Trimmer Architecture

Entity Tagger1 Trimmer Sentence

      

Sentence with Entity Tags

      

PERSON TIME EXPR

Compressions

                    

Parser2 Parse

1BBN IdentiFinder (Bikel et al., 1999) 2Charniak Parser (Charniak, 2000)

slide-9
SLIDE 9

9

Multi-candidate Trimmer

  • How to generate multiple candidate

compressions?

– Use the state of the parse tree after each rule application as a candidate – Use rules that generate multiple candidates – 9 single-output rules, 3 multi-output rules

  • Zajic et al, 2005, 2006; Zajic 2007
slide-10
SLIDE 10

10

Trimmer Rule: Root-S

  • Select node to be root of compression
  • Consider any S node with NP,VP children

The latest flood crest passed Chongqing in southwest China and waters were rising in Yichang on the middle reaches of the Yangtze state television reported Sunday S1 S S2 CC S3 NP VP

slide-11
SLIDE 11

11

Trimmer Rule: Conjunction

Illegal fireworks injured hundreds

  • f people

and started six fires S NP VP CC VP VP

  • Conjunction rule removes right, left or

neither child.

slide-12
SLIDE 12

12

Topiary Architecture

Topiary Sentence

      

Candidates

                    

Topic Assignment1 Document Document Corpus Topic Terms Compressions

                    

Trimmer

1BBN Unsupervised Topic

Detection

slide-13
SLIDE 13

13

Topiary Examples DUC2004

PINOCHET: wife appealed saying he too sick to be extradited to face charges MAHATHIR ANWAR_IBRAHIM: Lawyers went to court to demand client's release

– Mahathir Mohamad is the former Prime Minister of Malaysia – Anwar bin Ibrahim is a former deputy prime minister and finance minister of Malaysia, convicted of corruption in 1998

slide-14
SLIDE 14

14

Selector Architecture

Relevance & Centrality Scorer1 Sentence Selector Candidates + Features

                    

Document Document Set Candidates + More Features

                    

Query

?

Feature Weights Summary

     

Cull & Rescore

1Uniform Retrieval Architecture

(URA), UMD’s software infrastructure for IR tasks.

slide-15
SLIDE 15

15

Outline

  • Problem Description
  • MASC Architecture
  • MASC Results
  • Improving Candidate Selection
  • Summary & Future Work
slide-16
SLIDE 16

16

Evaluation of Headline Generation Systems

DUC2004 Test Data, Rouge recall with unigrams

0.16 0.18 0.2 0.22 0.24 0.26 0.28 0.3 First 75 UTD Topics HMM Hedge Trimmer Topiary HMM Hedge Trimmer Topiary

Rouge 1 Recall

No MASC MASC

slide-17
SLIDE 17

17

Evaluation of Multi-Document Summarization Systems

DUC2006 Test Data

0.05 0.055 0.06 0.065 0.07 0.075 No Compression HMM Hedge Trimmer Rouge 2 Recall

slide-18
SLIDE 18

18

Outline

  • Problem Description
  • MASC Architecture
  • MASC Results
  • Improving Candidate Selection
  • Summary & Future Work
slide-19
SLIDE 19

19

Tuning Feature Weights with ΔROUGE

c1 c2 ck

. . . Initialize: S = {}, H = {} C ← current k-best candidates for c ∈ C

ΔROUGE(c) = R2R(S∪{c}) - R2R(S) Add hypothesis to H

S ← S ∪ {c1} Update remaining candidates Repeat unless |S| > L wopt ← powellROUGE(H, w0)

Summary(S) Hypotheses(H) C

Δ1 Δ2 Δk

. . .

slide-20
SLIDE 20

20

Optimization Results

0.154 0.126 SU-4 0.104 0.081 2 0.403 0.363 1

ΔROUGE (k=10)

Manual

ROUGE

DUC2007 data, all differences significant at p < 0.05

Manual : Feature weights optimized manually to maximize ROUGE-2 Recall on the final system output Key Insights for ΔROUGE optimization:

  • Uses multiple alternative sentence compressions
  • Directly optimizes candidate selection process.
slide-21
SLIDE 21

21

  • Candidate words can be emitted by two disparate word

distributions

  • Assuming candidate words are i.i.d., the redundancy

feature for a given candidate is:

Redundancy

S = Summary, L = General English language

P(w | S) = n(w,S) S

( )

R(c) = log P(c)

( ) = log

P(w | S) + (1 )P(w | L)

wc

  • P(w | L) = n(w,L) L

( )

Other documents in the same cluster are used to represent the general language

REDUNDANT NON-REDUNDANT

λ + (1-λ)

P(w) =

slide-22
SLIDE 22

22

Incorporating Paraphrases

  • Redundancy uses bags-of-words to compute P(w|S)
  • Not useful if candidate word is a paraphrase of summary

word (classified as non-redundant)

  • Add another bag-of-words P, such that
  • Use n(w,P) for redundancy computation if n(w,S) = 0

P(w | S) = n(w,S) | S |

w S

P = { a paraphrase for w, }

slide-23
SLIDE 23

23

Generating Paraphrases

  • Leverage phrase-based MT system

– Use E-F correspondences extracted from word-aligned bi- text – Pivot each pair of E-F correspondence with common foreign side to get E-E correspondence –

  • Example
  • Pick most frequent correspondence for w

c(e1,e2) = c(e1, f )c( f ,e2)

f

  • 上升 ||| climbed ||| 1.0

上升 ||| increased ||| 2.0 上升 ||| uplifted ||| 1.0 increased ||| climbed ||| 2.0 climbed ||| uplifted ||| 1.0 . . . . . . uplifted ||| increased ||| 2.0

slide-24
SLIDE 24

24

Paraphrase Results

  • Using paraphrases yields no significant

improvements

  • Unrelated to the quality of the paraphrases
  • Anomalous cases occur extremely rarely

– The original bag-of-words is sufficient to capture candidate redundancy almost all the time

slide-25
SLIDE 25

25

Outline

  • Problem Description
  • MASC Architecture
  • MASC Results
  • Improving Candidate Selection
  • Summary & Future Work
slide-26
SLIDE 26

26

DUC 2007 Results

  • Systems 7, 36
  • Main:

– Responsiveness = 3.089 (4th) – ROUGE-2 = 0.108 (8th) – ROUGE-SU4 = 0.158 (11th)

  • Update:

– Responsiveness = 2.800 (2nd) – ROUGE-2 = 0.086 (9th) – ROUGE-SU4 = 0.124 (8th)

slide-27
SLIDE 27

27

Summary

  • MASC with feature-based candidate selection

improves headline generation and shows promise for multi-document summarization.

  • Optimizing for ΔROUGE provides significant

improvements over previous approach

  • Redundancy feature works at lexical as well as

document-level

  • Using paraphrases requires novel formulation
slide-28
SLIDE 28

28

Future Work

  • Fully explore Trimmer search space
  • Split redundancy feature into its components

and tune λ automatically

  • Use an n-gram LM to estimate P(w|L)
  • Continue to experiment with paraphrase-based

approaches to redundancy

– Scale up to phrase-level paraphrases – Use combination of high-coverage and high-quality paraphrases