Feuding Families and Former Friends: Unsupervised Learning for - - PowerPoint PPT Presentation

feuding families and former friends unsupervised learning
SMART_READER_LITE
LIVE PREVIEW

Feuding Families and Former Friends: Unsupervised Learning for - - PowerPoint PPT Presentation

Feuding Families and Former Friends: Unsupervised Learning for Dynamic Fictional Relationships Mohit Iyyer, Anupam Guha, Snigdha Chaturvedi, Jordan Boyd-Graber, and Hal Daum III University of Maryland, College Park University of


slide-1
SLIDE 1

Feuding Families and Former Friends:
 Unsupervised Learning for Dynamic Fictional Relationships

Mohit Iyyer, Anupam Guha, Snigdha Chaturvedi, Jordan Boyd-Graber, and Hal Daumé III

University of Maryland, College Park University of Colorado, Boulder

1

slide-2
SLIDE 2

How can we describe a fictional relationship between two characters?

2

slide-3
SLIDE 3

How can we describe a fictional relationship between two characters?

  • isn’t this easy? we can assign it a single label (or

relationship descriptor) from a predetermined set Friend or foe?

3

slide-4
SLIDE 4

How can we describe a fictional relationship between two characters?

  • isn’t this easy? we can assign it a single label (or

relationship descriptor) from a predetermined set Friend or foe?

Peter Pan and Captain Hook (Peter Pan)

4

slide-5
SLIDE 5

How can we describe a fictional relationship between two characters?

  • isn’t this easy? we can assign it a single label (or

relationship descriptor) from a predetermined set Friend or foe?

Peter Pan and Captain Hook (Peter Pan)

5

slide-6
SLIDE 6

How can we describe a fictional relationship between two characters?

  • isn’t this easy? we can assign it a single label (or

relationship descriptor) from a predetermined set Friend or foe?

Peter Pan and Captain Hook (Peter Pan) Frodo and Sam (Lord of the Rings)

6

slide-7
SLIDE 7

How can we describe a fictional relationship between two characters?

  • isn’t this easy? we can assign it a single label (or

relationship descriptor) from a predetermined set Friend or foe?

Peter Pan and Captain Hook (Peter Pan) Frodo and Sam (Lord of the Rings)

7

slide-8
SLIDE 8

How can we describe a fictional relationship between two characters?

  • isn’t this easy? we can assign it a single label (or

relationship descriptor) from a predetermined set Friend or foe?

Peter Pan and Captain Hook (Peter Pan) Frodo and Sam (Lord of the Rings) Winston and Julia (1984)

8

slide-9
SLIDE 9

How can we describe a fictional relationship between two characters?

  • isn’t this easy? we can assign it a single label (or

relationship descriptor) from a predetermined set Friend or foe?

Peter Pan and Captain Hook (Peter Pan) Frodo and Sam (Lord of the Rings) Winston and Julia (1984) ???

9

slide-10
SLIDE 10

How can we describe a fictional relationship between two characters?

  • isn’t this easy? we can assign it a single label (or

relationship descriptor) from a predetermined set Friend or foe?

Peter Pan and Captain Hook (Peter Pan) Frodo and Sam (Lord of the Rings) Winston and Julia (1984) ??? Harry Potter and Sirius (Prisoner of Azkaban)

10

slide-11
SLIDE 11

How can we describe a fictional relationship between two characters?

  • isn’t this easy? we can assign it a single label (or

relationship descriptor) from a predetermined set Friend or foe?

Peter Pan and Captain Hook (Peter Pan) Frodo and Sam (Lord of the Rings) Winston and Julia (1984) ??? Harry Potter and Sirius (Prisoner of Azkaban) ???

11

slide-12
SLIDE 12

How can we describe a fictional relationship between two characters?

  • what if we treat relationships as sequences (or

trajectories) of descriptors? (Chaturvedi et al., 2016) Tom Sawyer and Becky Thatcher: friends -> foes -> friends

12

slide-13
SLIDE 13

How can we describe a fictional relationship between two characters?

  • what if we treat relationships as sequences (or

trajectories) of descriptors? (Chaturvedi et al., 2016) Tom Sawyer and Becky Thatcher: friends -> foes -> friends

  • limited by fixed descriptor set
  • required expensive annotations
  • limited to plot summaries

13

slide-14
SLIDE 14

passage of time

Arthur and Lucy (Dracula)

slide-15
SLIDE 15

passage of time

I love him more than ever. We are to be married on 28 September.

Arthur and Lucy (Dracula)

slide-16
SLIDE 16

passage of time

love joy marriage

I love him more than ever. We are to be married on 28 September.

Arthur and Lucy (Dracula)

slide-17
SLIDE 17

passage of time

love joy marriage

I love him more than ever. We are to be married on 28 September.

I feel so weak and worn

  • ut … looked quite grieved

… I hadn't the spirit

Arthur and Lucy (Dracula)

slide-18
SLIDE 18

passage of time

love joy marriage love sadness sickness

I love him more than ever. We are to be married on 28 September.

I feel so weak and worn

  • ut … looked quite grieved

… I hadn't the spirit

Arthur and Lucy (Dracula)

slide-19
SLIDE 19

passage of time

love joy marriage love sadness sickness

I love him more than ever. We are to be married on 28 September.

I feel so weak and worn

  • ut … looked quite grieved

… I hadn't the spirit

poor girl, there is peace for her at

  • last. It is the end!

Arthur and Lucy (Dracula)

slide-20
SLIDE 20

passage of time

love joy marriage love sadness sickness love fantasy sickness sadness death

I love him more than ever. We are to be married on 28 September.

I feel so weak and worn

  • ut … looked quite grieved

… I hadn't the spirit

poor girl, there is peace for her at

  • last. It is the end!

Arthur and Lucy (Dracula)

slide-21
SLIDE 21

passage of time

love joy marriage love sadness sickness love fantasy sickness sadness death

I love him more than ever. We are to be married on 28 September.

I feel so weak and worn

  • ut … looked quite grieved

… I hadn't the spirit

poor girl, there is peace for her at

  • last. It is the end!

Arthur placed the stake over her heart … he struck with all his might. The Thing in the coffin writhed …

Arthur and Lucy (Dracula)

slide-22
SLIDE 22

passage of time

love joy marriage love sadness sickness love fantasy sickness sadness death love fantasy death murder

I love him more than ever. We are to be married on 28 September.

I feel so weak and worn

  • ut … looked quite grieved

… I hadn't the spirit

poor girl, there is peace for her at

  • last. It is the end!

Arthur placed the stake over her heart … he struck with all his might. The Thing in the coffin writhed …

Arthur and Lucy (Dracula)

slide-23
SLIDE 23

Why is this a worthwhile problem?

  • “Distant reading” (Moretti, 2005) can help humanities

scholars collect examples of specific relationship types

15

“Do Jane Austen’s female and male protagonists have a pattern in their evolving relationship (e.g., mutual disdain followed by romantic love)?” (Butler, 1975; Stovel, 1987; Hinant, 2006) “Do certain authors or novels portray relationships of desire more than others?” (Polhemus, 1990) “Can we detect positive or negative subtext underlying meals between two characters?” (Foster, 2009; Cognard-Black et al., 2014)

slide-24
SLIDE 24

Outline

  • Dataset: character interactions
  • RMN: relationship modeling network
  • Experiments: coherent descriptors, interpretable

trajectories

  • Analysis: RMN’s strengths and weaknesses

16

slide-25
SLIDE 25

A Dataset of Character Interactions

17

  • For each pair of characters in a particular book, we extract

all spans of text that contain mentions to both characters

slide-26
SLIDE 26

A Dataset of Character Interactions

17

“If anyone was ever minding his business, it was I," Ignatius breathed. "Please. We must stop. I think I'm going to have a hemorrhage.” "Okay." Mrs. Reilly looked at her son's reddening face and realized that he would very happily collapse at her feet just to prove his point.”

t=0

  • For each pair of characters in a particular book, we extract

all spans of text that contain mentions to both characters

slide-27
SLIDE 27

A Dataset of Character Interactions

17

“If anyone was ever minding his business, it was I," Ignatius breathed. "Please. We must stop. I think I'm going to have a hemorrhage.” "Okay." Mrs. Reilly looked at her son's reddening face and realized that he would very happily collapse at her feet just to prove his point.”

t=0

“Ignatius belched the gas of a dozen brownies trapped by his valve. "Grant me a little peace.…” "You know I appreciate you, babe," Mrs. Reilly sniffed. "Come on and gimme a little goodbye kiss like a good boy.”

t=1

  • For each pair of characters in a particular book, we extract

all spans of text that contain mentions to both characters

slide-28
SLIDE 28

A Dataset of Character Interactions

17

“If anyone was ever minding his business, it was I," Ignatius breathed. "Please. We must stop. I think I'm going to have a hemorrhage.” "Okay." Mrs. Reilly looked at her son's reddening face and realized that he would very happily collapse at her feet just to prove his point.”

t=0

“Ignatius belched the gas of a dozen brownies trapped by his valve. "Grant me a little peace.…” "You know I appreciate you, babe," Mrs. Reilly sniffed. "Come on and gimme a little goodbye kiss like a good boy.”

t=1

  • Mrs. Reilly looked at her son slyly and asked, "Ignatius, you sure you

not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling

  • building. No!"

t=2

  • For each pair of characters in a particular book, we extract

all spans of text that contain mentions to both characters

slide-29
SLIDE 29

A Dataset of Character Interactions

  • 1,383 novels from Project Gutenberg and other

Internet sources

  • Genres represented include romance, mystery, and fantasy
  • Preprocessed with David Bamman’s BookNLP pipeline
  • Each span is a 200-token window centered around a

character mention

  • 20,013 unique character pairs and 380,408 spans

18

slide-30
SLIDE 30
  • recurrent autoencoder with dictionary learning

Relationship Modeling Network (RMN)

19

slide-31
SLIDE 31
  • recurrent autoencoder with dictionary learning

Relationship Modeling Network (RMN)

20

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!" “Ignatius belched the gas of a dozen brownies trapped by his valve. "Grant me a little peace.…” "You know I appreciate you, babe," Mrs. Reilly

  • sniffed. "Come on and gimme a little goodbye

kiss like a good boy.”

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

slide-32
SLIDE 32
  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!" “Ignatius belched the gas of a dozen brownies trapped by his valve. "Grant me a little peace.…” "You know I appreciate you, babe," Mrs. Reilly

  • sniffed. "Come on and gimme a little goodbye

kiss like a good boy.”

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

  • recurrent autoencoder with dictionary learning

Relationship Modeling Network (RMN)

21

reconstruct inputs

slide-33
SLIDE 33
  • recurrent autoencoder with dictionary learning
  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!" “Ignatius belched the gas of a dozen brownies trapped by his valve. "Grant me a little peace.…” "You know I appreciate you, babe," Mrs. Reilly

  • sniffed. "Come on and gimme a little goodbye

kiss like a good boy.”

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

Relationship Modeling Network (RMN)

22

descriptor matrix

slide-34
SLIDE 34
  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!" “Ignatius belched the gas of a dozen brownies trapped by his valve. "Grant me a little peace.…” "You know I appreciate you, babe," Mrs. Reilly

  • sniffed. "Come on and gimme a little goodbye

kiss like a good boy.”

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

  • recurrent autoencoder with dictionary learning

Relationship Modeling Network (RMN)

23

reconstruct inputs

slide-35
SLIDE 35

24

rt = RTdt

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

  • Mrs. Reilly

Ignatius “A Confederacy

  • f Dunces”

ht = f(Wh · [vst; vc1; vc2; vb])

vst vc1 vc2 vb dt−1 R dt = α · softmax(Wd · [ht; dt−1])+ (1 − α) · dt−1

: previous state : descriptor matrix : reconstruction

  • f input span

: distribution over descriptors

  • 6. make the

reconstructed vector close to the input span vector

  • 1. word

embedding average

  • 2. mix with

character embeddings and book embeddings

  • 3. a recurrent

connection that copies over some

  • f the previous

hidden state

  • 4. use a softmax

activation function to make the hidden state a probability distribution over descriptors…

  • 5. multiply the hidden

state by the descriptor matrix to obtain a reconstruction of the span vector

slide-36
SLIDE 36

25

rt = RTdt

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

  • Mrs. Reilly

Ignatius “A Confederacy

  • f Dunces”

ht = f(Wh · [vst; vc1; vc2; vb])

vst vc1 vc2 vb dt−1 R dt = α · softmax(Wd · [ht; dt−1])+ (1 − α) · dt−1

: previous state : descriptor matrix : reconstruction

  • f input span

: distribution over descriptors

  • 6. make the

reconstructed vector close to the input span vector

  • 1. word

embedding average

  • 2. mix with

embeddings for characters and books

  • 3. a recurrent

connection that copies over some

  • f the previous

hidden state

  • 4. use a softmax

activation function to make the hidden state a probability distribution over descriptors…

  • 5. multiply the hidden

state by the descriptor matrix to obtain a reconstruction of the span vector

slide-37
SLIDE 37

26

rt = RTdt

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

  • Mrs. Reilly

Ignatius “A Confederacy

  • f Dunces”

ht = f(Wh · [vst; vc1; vc2; vb])

vst vc1 vc2 vb dt−1 R dt = α · softmax(Wd · [ht; dt−1])+ (1 − α) · dt−1

: previous state : descriptor matrix : reconstruction

  • f input span

: distribution over descriptors

  • 6. make the

reconstructed vector close to the input span vector

  • 1. word

embedding average

  • 2. mix with

embeddings for characters and books

  • 3. a recurrent

connection that copies over some

  • f the previous

hidden state

  • 4. use a softmax

activation function to make the hidden state a probability distribution over descriptors…

  • 5. multiply the hidden

state by the descriptor matrix to obtain a reconstruction of the span vector

slide-38
SLIDE 38

27

rt = RTdt

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

  • Mrs. Reilly

Ignatius “A Confederacy

  • f Dunces”

ht = f(Wh · [vst; vc1; vc2; vb])

vst vc1 vc2 vb dt−1 R dt = α · softmax(Wd · [ht; dt−1])+ (1 − α) · dt−1

: previous state : descriptor matrix : reconstruction

  • f input span

: distribution over descriptors

  • 6. make the

reconstructed vector close to the input span vector

  • 1. word

embedding average

  • 2. mix with

embeddings for characters and books

  • 3. a recurrent

connection that copies over some

  • f the previous

hidden state

  • 4. use a softmax

activation function to make the hidden state a probability distribution over descriptors…

  • 5. multiply the hidden

state by the descriptor matrix to obtain a reconstruction of the span vector

linear interpolation

slide-39
SLIDE 39

28

rt = RTdt

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

  • Mrs. Reilly

Ignatius “A Confederacy

  • f Dunces”

ht = f(Wh · [vst; vc1; vc2; vb])

vst vc1 vc2 vb dt−1 R dt = α · softmax(Wd · [ht; dt−1])+ (1 − α) · dt−1

: previous state : descriptor matrix : reconstruction

  • f input span

: distribution over descriptors

  • 6. make the

reconstructed vector close to the input span vector

  • 1. word

embedding average

  • 2. mix with

embeddings for characters and books

  • 3. a recurrent

connection that copies over some

  • f the previous

hidden state

  • 4. use a softmax

activation function to make the hidden state a probability distribution over descriptors…

  • 5. multiply the hidden

state by the descriptor matrix to obtain a reconstruction of the span vector

softmax activation

slide-40
SLIDE 40

29

rt = RTdt

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

  • Mrs. Reilly

Ignatius “A Confederacy

  • f Dunces”

ht = f(Wh · [vst; vc1; vc2; vb])

vst vc1 vc2 vb dt−1 R dt = α · softmax(Wd · [ht; dt−1])+ (1 − α) · dt−1

: previous state : descriptor matrix : reconstruction

  • f input span

: distribution over descriptors

  • 6. make the

reconstructed vector close to the input span vector

  • 1. word

embedding average

  • 2. mix with

embeddings for characters and books

  • 3. a recurrent

connection that copies over some

  • f the previous

hidden state

  • 4. use a softmax

activation function to make the hidden state a probability distribution over descriptors…

  • 5. multiply the hidden

state by the descriptor matrix to obtain a reconstruction of the span vector

slide-41
SLIDE 41

30

rt = RTdt

  • Mrs. Reilly looked at her son slyly and asked,

"Ignatius, you sure you not a communiss?" "Oh, my God!" Ignatius bellowed. "Every day I am subjected to a McCarthyite witchhunt in this crumbling building. No!"

  • Mrs. Reilly

Ignatius “A Confederacy

  • f Dunces”

ht = f(Wh · [vst; vc1; vc2; vb])

vst vc1 vc2 vb dt−1 R dt = α · softmax(Wd · [ht; dt−1])+ (1 − α) · dt−1

: previous state : descriptor matrix : reconstruction

  • f input span

: distribution over descriptors

  • 6. make the

reconstructed vector close to the input span vector

  • 1. word

embedding average

  • 2. mix with

embeddings for characters and books

  • 3. a recurrent

connection that copies over some

  • f the previous

hidden state

  • 4. use a softmax

activation function to make the hidden state a probability distribution over descriptors…

  • 5. multiply the hidden

state by the descriptor matrix to obtain a reconstruction of the span vector

slide-42
SLIDE 42

Labeling the Learned Descriptors

  • We compute the nearest word embeddings to each

row of the descriptor matrix R, which humans use to provide external labels.

31

violence: grenades, guns, bullets sadness: regretful, rueful, pity politics: political, leadership, rule fantasy: cosmic, realms, universe suffering: fear, nightmares, suffer

slide-43
SLIDE 43

Relationship to Topic Models

  • RMN outputs ≈ topic model latent variables:
  • descriptor matrix R ≈ topic-word matrices ϕ
  • descriptor weights dt at each timestep ≈

document-topic assignments z

  • Baselines:
  • temporally-oblivious: LDA (Blei et al., 2001), Nubbi (Chang et al.,

2008)

  • temporally-aware: HTMM (Gruber et al., 2007)

32

slide-44
SLIDE 44

Experiment 1:

Descriptor Coherence

33

slide-45
SLIDE 45

Do the Descriptors Make Sense?

  • Goal: compare the descriptors learned by the RMN to

the topics learned by our topic model baselines

  • Task: word intrusion (Chang et al., 2009)
  • Workers identify an “intruder” word from a set of words that

—other than the intruder—come from the same descriptor

34

contempt malice condescend praise distaste mock worship pray devote yourselves gods gather

slide-46
SLIDE 46

Do the Descriptors Make Sense?

  • Goal: compare the descriptors learned by the RMN to

the topics learned by our topic model baselines

  • Task: word intrusion (Chang et al., 2009)
  • Workers identify an “intruder” word from a set of words that

—other than the intruder—come from the same descriptor

35

contempt malice condescend praise distaste mock worship pray devote yourselves gods gather

slide-47
SLIDE 47

Do the Descriptors Make Sense?

  • Goal: compare the descriptors learned by the RMN to

the topics learned by our topic model baselines

  • Task: word intrusion (Chang et al., 2009)
  • Workers identify an “intruder” word from a set of words that

—other than the intruder—come from the same descriptor

36

contempt malice condescend praise distaste mock worship pray devote yourselves gods gather

slide-48
SLIDE 48

Do the Descriptors Make Sense?

37

slide-49
SLIDE 49

K=10 K=30 K=50 0.4 0.5 0.6 0.7 0.8

Model Precision

LDA Nubbi HTMM GRMN RMN

Do the Descriptors Make Sense?

38

slide-50
SLIDE 50

Coherent Descriptors

39

HTMM crime: blood knife pain legs steal food: kitchen mouth glass food bread violence: sword shot blood shouted swung boats: ship boat captain deck crew

  • utdoors: stone rock path darkness desert

RMN

  • utdoors: outdoors trail trails hillside grassy slopes

sadness: regretful rueful pity pained despondent education: teaching graduate year teacher attended love: love delightful happiness enjoyed enjoyable murder: autopsy arrested homicide murdered

slide-51
SLIDE 51

Coherent Descriptors

40

HTMM crime: blood knife pain legs steal food: kitchen mouth glass food bread violence: sword shot blood shouted swung boats: ship boat captain deck crew

  • utdoors: stone rock path darkness desert

RMN

  • utdoors: outdoors trail trails hillside grassy slopes

sadness: regretful rueful pity pained despondent education: teaching graduate year teacher attended love: love delightful happiness enjoyed enjoyable murder: autopsy arrested homicide murdered

slide-52
SLIDE 52

Experiment 2:

Trajectory Quality

41

slide-53
SLIDE 53

Visualizing Trajectories

  • for all time steps t,

compute argmax

  • f dt and stack

vertically

42

slide-54
SLIDE 54
  • for all time steps t,

compute argmax

  • f dt and stack

vertically

Visualizing Trajectories

43

time love death money crime 0.95 0.01 0.03 0.01 1 0.8 0.01 0.18 0.01 2 0.4 0.01 0.5 0.09 3 0.3 0.01 0.2 0.5 4 0.2 0.7 0.05 0.05

slide-55
SLIDE 55

Visualizing Trajectories

44

love time

time love death money crime 0.95 0.01 0.03 0.01 1 0.8 0.01 0.18 0.01 2 0.4 0.01 0.5 0.09 3 0.3 0.01 0.2 0.5 4 0.2 0.7 0.05 0.05

  • for all time steps t,

compute argmax

  • f dt and stack

vertically

slide-56
SLIDE 56

Visualizing Trajectories

45

love time

time love death money crime 0.95 0.01 0.03 0.01 1 0.8 0.01 0.18 0.01 2 0.4 0.01 0.5 0.09 3 0.3 0.01 0.2 0.5 4 0.2 0.7 0.05 0.05

  • for all time steps t,

compute argmax

  • f dt and stack

vertically

slide-57
SLIDE 57

Visualizing Trajectories

46

money love time

time love death money crime 0.95 0.01 0.03 0.01 1 0.8 0.01 0.18 0.01 2 0.4 0.01 0.5 0.09 3 0.3 0.01 0.2 0.5 4 0.2 0.7 0.05 0.05

  • for all time steps t,

compute argmax

  • f dt and stack

vertically

slide-58
SLIDE 58

Visualizing Trajectories

47

money love time

time love death money crime 0.95 0.01 0.03 0.01 1 0.8 0.01 0.18 0.01 2 0.4 0.01 0.5 0.09 3 0.3 0.01 0.2 0.5 4 0.2 0.7 0.05 0.05

crime

  • for all time steps t,

compute argmax

  • f dt and stack

vertically

slide-59
SLIDE 59

Visualizing Trajectories

48

death money love time

time love death money crime 0.95 0.01 0.03 0.01 1 0.8 0.01 0.18 0.01 2 0.4 0.01 0.5 0.09 3 0.3 0.01 0.2 0.5 4 0.2 0.7 0.05 0.05

crime

  • for all time steps t,

compute argmax

  • f dt and stack

vertically

slide-60
SLIDE 60

Do the Trajectories Make Sense?

  • We crawl Wikipedia and SparkNotes for summaries
  • Removing uninformative summaries results in 125 character

pairs to evaluate

  • Workers prefer the RMN to the HTMM for 87 out of the 125

relationships (69.6%, Fleiss κ=0.32)

49

In this task, you will be comparing two timelines

  • f how a relationship between a pair of literary

characters changes over time. We will provide you with a summary of the relationship, and your job is to select which of the two timelines (A or B) better captures the content of the summary.

slide-61
SLIDE 61

50

Summary: Govinda is Siddhartha’s best friend and sometimes his follower. Like Siddhartha, Govinda devotes his life to the quest for understanding and

  • enlightenment. He leaves his village with Siddhartha to join the Samanas, then

leaves the Samanas to follow Gotama. He searches for enlightenment independently

  • f Siddhartha but persists in looking for teachers who can show him the way. In the

end, he is able to achieve enlightenment only because of Siddhartha’s love for him.

slide-62
SLIDE 62

Qualitative Analysis:

Good and Bad Trajectories

51

slide-63
SLIDE 63

52

TIME HTMM RMN Dracula: Arthur and Lucy

love love sadness joy love fantasy love fantasy sickness death sadness death marriage sickness murder

passage of time

I love him more than ever. We are to be married on 28 September. I feel so weak and worn

  • ut … looked quite grieved

… I hadn't the spirit poor girl, there is peace for her at

  • last. It is the end!

Arthur placed the stake over her heart … he struck with all his might. The Thing in the coffin writhed …

Arthur and Lucy “ground-truth”:

marriage -> sickness -> death -> murder

slide-64
SLIDE 64

53

TIME HTMM RMN Dracula: Arthur and Lucy

love love sadness joy love fantasy love fantasy sickness death sadness death marriage sickness murder

passage of time

I love him more than ever. We are to be married on 28 September. I feel so weak and worn

  • ut … looked quite grieved

… I hadn't the spirit poor girl, there is peace for her at

  • last. It is the end!

Arthur placed the stake over her heart … he struck with all his might. The Thing in the coffin writhed …

Arthur and Lucy “ground-truth”:

marriage -> sickness -> death -> murder

learned trajectories:

slide-65
SLIDE 65

54

HTMM RMN Storm Island: David and Lucy HTMM RMN A Tale of Two Cities: Darnay and Lucie

Event-based similarities between the two models The RMN is led astray by the novel’s sad tone

slide-66
SLIDE 66

55

HTMM RMN Storm Island: David and Lucy HTMM RMN A Tale of Two Cities: Darnay and Lucie

Event-based similarities between the two models The RMN is led astray by the novel’s sad tone

slide-67
SLIDE 67

Qualitative Analysis:

Using Existing Datasets

56

slide-68
SLIDE 68

What Makes a Good Relationship?

  • Dataset of Massey et al. (2015) has affinity

annotations for relationships in Project Gutenberg

  • 120 non-neutral relationships are also present in our dataset

57

slide-69
SLIDE 69

What Makes a Good Relationship?

  • Dataset of Massey et al. (2015) has affinity

annotations for relationships in Project Gutenberg

  • 120 non-neutral relationships are also present in our dataset

58

love death sadness 0.9 0.05 0.05 0.8 0.1 0.1 0.6 0.3 0.1 0.7 0.1 0.2 0.8 0.1 0.1 love death sadness 0.1 0.7 0.2 0.2 0.3 0.5 0.15 0.25 0.6 0.05 0.65 0.3 0.1 0.2 0.7

positive negative

Don Quixote & Sancho Panza Candide & Cunégonde Anna Karenina & Vronsky … Dracula & Jonathan Harker

  • Dr. Jekyll & Mr. Hyde

Hester Prynne & Chillingworth …

slide-70
SLIDE 70

What Makes a Good Relationship?

  • Dataset of Massey et al. (2015) has affinity

annotations for relationships in Project Gutenberg

  • 120 non-neutral relationships are also present in our dataset

59

love death sadness 0.9 0.05 0.05 0.8 0.1 0.1 0.6 0.3 0.1 0.7 0.1 0.2 0.8 0.1 0.1 0.76 0.13 0.11 love death sadness 0.1 0.7 0.2 0.2 0.3 0.5 0.15 0.25 0.6 0.05 0.65 0.3 0.1 0.2 0.7 0.12 0.42 0.46

positive negative average the positive and negative trajectories

slide-71
SLIDE 71

What Makes a Good Relationship?

  • Dataset of Massey et al. (2015) has affinity

annotations for relationships in Project Gutenberg

  • 120 non-neutral relationships are also present in our dataset

60

love death sadness 0.9 0.05 0.05 0.8 0.1 0.1 0.6 0.3 0.1 0.7 0.1 0.2 0.8 0.1 0.1 0.76 0.13 0.11 love death sadness 0.1 0.7 0.2 0.2 0.3 0.5 0.15 0.25 0.6 0.05 0.65 0.3 0.1 0.2 0.7 0.12 0.42 0.46

positive negative

  • 1. love
  • 2. death
  • 3. sadness
  • 1. sadness
  • 2. death
  • 3. love
slide-72
SLIDE 72

61

RMN education love religion sex HTMM love parental business

  • utdoors

Most Positive Descriptors

slide-73
SLIDE 73

62

RMN education love religion sex RMN politics murder sadness royalty HTMM love parental business

  • utdoors

HTMM love politics violence crime Most Positive Descriptors Most Negative Descriptors

slide-74
SLIDE 74

63

RMN education love religion sex RMN politics murder sadness royalty HTMM love parental business

  • utdoors

HTMM love politics violence crime Most Positive Descriptors Most Negative Descriptors

slide-75
SLIDE 75

Why is Politics Negative?

  • Both models rank politics as highly negative
  • The affinity data we look at comes primarily from

Victorian-era authors (e.g., Charles Dickens and George Eliot)

64

Victorian-era authors are “obsessed with otherness… of antiquated social and legal institutions, and of autocratic and/or dictatorial abusive government” (Zarifopol-Johnston,1995)

slide-76
SLIDE 76

Areas for Improvement

  • Difficult to evaluate unsupervised relationship

modeling, requires considerable human effort

  • Our data processing leaves out a lot of information
  • e.g., spans of text in which one but not both characters in a

relationship are mentioned

  • only considers undirected relationships between pairs
  • Model performance is directly tied to the quality of

character disambiguation and coreference resolution

  • e.g., first person pronouns

65

slide-77
SLIDE 77

Recap

  • Introduced the task of unsupervised relationship

modeling as well as an interpretable neural network architecture, the RMN, for this task

  • Found that the RMN generates higher quality

descriptors and more interpretable trajectories than topic model baselines

  • Future work: collaborate with humanities researchers

to help answer literary questions with the RMN

66

slide-78
SLIDE 78

Thanks! Questions?

code/data @ github.com/miyyer/rmn

67