Public Self-consciousness for Endowing Dialogue Agents with - - PowerPoint PPT Presentation

public self consciousness for endowing dialogue agents
SMART_READER_LITE
LIVE PREVIEW

Public Self-consciousness for Endowing Dialogue Agents with - - PowerPoint PPT Presentation

Public Self-consciousness for Endowing Dialogue Agents with Consistent Persona 2020 BAICS workshop (Oral) Hyunwoo Kim Byeongchang Kim Gunhee Kim V I S I O N & L E A R N I N G L A B SEOUL NATIONAL UNIVERSITY The Consistency Problem


slide-1
SLIDE 1

Public Self-consciousness for Endowing Dialogue Agents with Consistent Persona

2020 BAICS workshop (Oral) Hyunwoo Kim Byeongchang Kim Gunhee Kim

SEOUL NATIONAL UNIVERSITY V I S I O N & L E A R N I N G L A B

slide-2
SLIDE 2

The Consistency Problem

Bot: I’m a programmer. Human: What is your job? Human: What do you do? Bot: I’m a lawyer. Human: ???

in Dialogue Agents

slide-3
SLIDE 3

Previous works tackling the Consistency Problem

Embeddings Benchmark Datasets Natural Language Inference (NLI)

slide-4
SLIDE 4

Dolan et al. 2016. A persona-based neural conversational model. ACL

  • Feed a persona embedding to the decoder along with the target utterance

Previous Works:

Input persona embeddings to the model

slide-5
SLIDE 5

A dialogue dataset involving two interlocutors getting to know each other while playing the given persona

  • the PersonaChat dataset

Zhang et al. 2018. Personalizing Dialogue Agents: I have a dog, do you have pets too? ACL

Previous Works:

Benchmark dataset which persona sentences are given to the model

slide-6
SLIDE 6

Previous Works:

Exploit Natural Language Inference (NLI) annotations

Given a “premise”, the task of determining whether a “hypothesis” is

  • True (Entailment)
  • False (Contradiction)
  • Undetermined (Neutral)

Premise: I love to go for a drive with my new car.

  • Hypothesis: Recently, I finally bought a car!
  • Hypothesis: I do not have a car.
  • Hypothesis: Milk shake is my favorite dessert.

[Entailment] [Contradiction] [Neutral]

Welleck et al. 2019. Dialogue Natural Language Inference. ACL

slide-7
SLIDE 7
  • 1. collect additional NLI annotations

Previous Works: use NLI

Welleck et al. 2019. Dialogue Natural Language Inference. ACL

slide-8
SLIDE 8
  • 2. train external NLI model on the annotation

Chen et al. 2017. Enhanced LSTM for Natural Language Inference. EMNLP (left) Conneau et al. 2017. Supervised Learning of Universal Sentence Representations from Natural Language Inference Data. ACL (right)

Previous Works: use NLI

slide-9
SLIDE 9
  • 3. compute pair-wise contradiction scores on

every candidate sentences of the dialogue agent and persona sentences to re-weight contradicting candidates

Persona sentence 1 Persona sentence 2 … Persona sentence 4 Persona sentence m Candidate sentence 1 Candidate sentence 2 Candidate sentence 3 … Candidate sentence 8 Candidate sentence 9 Candidate sentence n

Welleck et al. 2019. Dialogue Natural Language Inference. ACL Song et al. 2019. Generating Persona Consistent Dialogues by Exploiting Natural Language Inference. arXiv

Previous Works: use NLI

Compute contradiction score with NLI model for each pair

slide-10
SLIDE 10

Limitations

1. Require NLI annotations on the target dataset 2. Require training external NLI model on the annotations 3. NLI model computes pair-wise contradiction score for every persona sentences and candidate sentences

Demanding & Inscalable Previous Works: use NLI

slide-11
SLIDE 11

Our question:

How do humans maintain consistency?

slide-12
SLIDE 12

We do not ask others whether we are consistent or not

We ask ourselves.

slide-13
SLIDE 13

by predicting how we will be perceived by others

We ask ourselves.

slide-14
SLIDE 14

Public Self-Consciousness

The awareness of the self as a social object that can be observed and evaluated by others

slide-15
SLIDE 15

We model the self-consciousness through an imaginary listener

slide-16
SLIDE 16

Modeling a Listener:

The Bayesian Rational Speech Acts framework

Treats language use as a recursive process where probabilistic speaker and listener reason about each other in Bayesian fashion

Frank and Goodman. 2012. Predicting Pragmatic Reasoning in Language Games. Science

slide-17
SLIDE 17

Our approach: A self-conscious agent

thinking about how it will be perceived

slide-18
SLIDE 18

The Self-Conscious Speaker 𝑻𝟐

Base Speaker: 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

Imaginary Listener: 𝑀!

" (𝑗|𝑣$", ℎ, 𝑞")

Self-Conscious Speaker: 𝑇%

"

Persona: 𝑗 Dialogue History: ℎ

Learned Distractor Personas: 𝑗′

Speaker’s Utterance: 𝑣" ∝ 𝑀!

" 𝑗 ℎ, 𝑣$", 𝑞" &

∗ 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

𝑞"'%(𝑗)

slide-19
SLIDE 19

’s Persona (Speaker 1’s Persona) I live in Florida and have a dog. I am going to college next year. I enjoy going outside and playing with my friends. I love Disney movies and animations. [Speaker 2] Hello, how are you today? [Speaker 1] Great! Just watching my favorite TV show. You? [Speaker 2] Cool! What do you like to do when COVID’s over?

𝑗: 𝑕𝑗𝑤𝑓𝑜 𝑞𝑓𝑠𝑡𝑝𝑜𝑏 ℎ: 𝑒𝑗𝑏𝑚𝑝𝑕𝑣𝑓 ℎ𝑗𝑡𝑢𝑝𝑠𝑧

Task Setting:

[Model’s generation]: 𝑣-, 𝑣., 𝑣/, … , 𝑣01-, 𝑣0 𝑣: 𝑣𝑢𝑢𝑓𝑠𝑏𝑜𝑑𝑓 (𝑢 𝑢𝑝𝑙𝑓𝑜𝑡)

slide-20
SLIDE 20

Intuitive Explanation of the Self-Conscious Speaker 𝑻𝟐

Self-Conscious Speaker

‘I want to be identified as my persona, not some other different persona.’

Distractors

’s Persona I live in Florida and I have a dog. I am going to college next year. I enjoy going outside to play. I love Disney movies and animations. ’s Persona I like reading books. I raise two cats. My girlfriend is a developer. I like to eat pepperoni pizza. ’s Persona I live in a big city I work at the gym as a trainer. I have two dogs. I like to watch extreme sports.

‘Will I sound like me?’

slide-21
SLIDE 21

I like to [ read books at the library ]

Intuitive Explanation of the Self-Conscious Speaker 𝑻𝟐

Self-Conscious Speaker

‘I want to be identified as my persona, not some other different persona.’

Distractors

’s Persona I live in Florida and I have a dog. I am going to college next year. I enjoy going outside to play. I love Disney movies and animations. ’s Persona I like reading books. I raise two cats. My girlfriend is a developer. I like to eat pepperoni pizza. ’s Persona I live in a big city I work at the gym as a trainer. I have two dogs. I like to watch extreme sports.

I like to ‘Will I sound like me?’

slide-22
SLIDE 22

I like to [ go to Disney World ] I like to [ read books at the library ]

Intuitive Explanation of the Self-Conscious Speaker 𝑻𝟐

Self-Conscious Speaker

‘I want to be identified as my persona, not some other different persona.’

Distractors

’s Persona I live in Florida and I have a dog. I am going to college next year. I enjoy going outside to play. I love Disney movies and animations. ’s Persona I like reading books. I raise two cats. My girlfriend is a developer. I like to eat pepperoni pizza. ’s Persona I live in a big city I work at the gym as a trainer. I have two dogs. I like to watch extreme sports.

‘Will I sound like me?’

slide-23
SLIDE 23

Intuitive Explanation of the Self-Conscious Speaker 𝑻𝟐

Self-Conscious Speaker

‘I want to be identified as my persona, not some other different persona.’

Distractors

’s Persona I live in Florida and I have a dog. I am going to college next year. I enjoy going outside to play. I love Disney movies and animations. ’s Persona I like reading books. I raise two cats. My girlfriend is a developer. I like to eat pepperoni pizza. ’s Persona I live in a big city I work at the gym as a trainer. I have two parrots. I like to watch extreme sports.

‘Will I sound like me?’ I like to [ go to Disney World ]

slide-24
SLIDE 24

Intuitive Explanation of the Self-Conscious Speaker 𝑻𝟐

Self-Conscious Speaker

‘I want to be identified as my persona, not some other different persona.’

Distractors

’s Persona I live in Florida and I have a dog. I am going to college next year. I enjoy going outside to play. I love Disney movies and animations.

‘Will I sound like me?’ I like to [ go to Disney World ]

Base Speaker: 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

Imaginary Listener: 𝑀!

" (𝑗|𝑣$", ℎ, 𝑞")

Self-Conscious Speaker: 𝑇%

"

Persona: 𝑗 Dialogue History: ℎ

Learned Distractor Personas: 𝑗′

Speaker’s Utterance: 𝑣" ∝ 𝑀!

" 𝑗 ℎ, 𝑣$", 𝑞" &

∗ 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

𝑞!"#(𝑗)

slide-25
SLIDE 25

Components of the Self-Conscious Speaker 𝑻𝟐

  • 𝐵 𝑐𝑏𝑡𝑓 𝑡𝑞𝑓𝑏𝑙𝑓𝑠

𝑇#

$ 𝑣$

𝑗, ℎ, 𝑣%$)

(𝑜𝑝 𝑡𝑓𝑚𝑔 𝑑𝑝𝑜𝑡𝑑𝑗𝑝𝑣𝑡𝑜𝑓𝑡𝑡)

  • 𝑈ℎ𝑓 𝒕𝒇𝒎𝒈 𝒅𝒑𝒐𝒕𝒅𝒋𝒑𝒗𝒕 𝑡𝑞𝑓𝑏𝑙𝑓𝑠

𝑇&

$ 𝑣$

𝑗, ℎ, 𝑣%$) ∝ 𝑀#

$ 𝑗 ℎ, 𝑣'$, 𝑞$)( * 𝑇# $ 𝑣$ 𝑗, ℎ, 𝑣%$)

  • 𝐵𝑜 𝑗𝑛𝑏𝑕𝑗𝑜𝑏𝑠𝑧 𝑚𝑗𝑡𝑢𝑓𝑜𝑓𝑠

𝑀!

" 𝑗 ℎ, 𝑣$", 𝑞") ∝

𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")* J 𝑞"(𝑗)

∑+'∈- 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")* J 𝑞" (𝑗.)

A Recursive Process in Bayesian Fashion

Base Speaker: 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

Imaginary Listener: 𝑀!

" (𝑗|𝑣$", ℎ, 𝑞")

Self-Conscious Speaker: 𝑇%

"

Persona: 𝑗 Dialogue History: ℎ

Learned Distractor Personas: 𝑗′

Speaker’s Utterance: 𝑣" ∝ 𝑀!

" 𝑗 ℎ, 𝑣$", 𝑞" &

∗ 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

𝑞!"#(𝑗)

slide-26
SLIDE 26

Base Speaker 𝑻𝟏

  • 𝐵 𝑐𝑏𝑡𝑓 𝑡𝑞𝑓𝑏𝑙𝑓𝑠

𝑇#

$ 𝑣$

𝑗, ℎ, 𝑣%$)

(𝑜𝑝 𝑡𝑓𝑚𝑔 𝑑𝑝𝑜𝑡𝑑𝑗𝑝𝑣𝑡𝑜𝑓𝑡𝑡)

Any pretrained generative dialogue model = Prior distribution

Base Speaker: 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

Imaginary Listener: 𝑀!

" (𝑗|𝑣$", ℎ, 𝑞")

Self-Conscious Speaker: 𝑇%

"

Persona: 𝑗 Dialogue History: ℎ

Learned Distractor Personas: 𝑗′

Speaker’s Utterance: 𝑣" ∝ 𝑀!

" 𝑗 ℎ, 𝑣$", 𝑞" &

∗ 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

𝑞!"#(𝑗)

Generating one token at a time

slide-27
SLIDE 27

Imaginary Listener 𝑴𝟏

  • 𝐵𝑜 𝑗𝑛𝑏𝑕𝑗𝑜𝑏𝑠𝑧 𝑚𝑗𝑡𝑢𝑓𝑜𝑓𝑠

𝑀!

" 𝑗 ℎ, 𝑣$", 𝑞") ∝

𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")* J 𝑞"(𝑗)

∑+'∈- 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")* J 𝑞" (𝑗.)

Base Speaker: 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

Imaginary Listener: 𝑀!

" (𝑗|𝑣$", ℎ, 𝑞")

Self-Conscious Speaker: 𝑇%

"

Persona: 𝑗 Dialogue History: ℎ

Learned Distractor Personas: 𝑗′

Speaker’s Utterance: 𝑣" ∝ 𝑀!

" 𝑗 ℎ, 𝑣$", 𝑞" &

∗ 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

𝑞!"#(𝑗)

The likelihood of the given persona World 𝑱: given persona + distractors Accumulative World Prior

Learned with Life-long Memory Networks

Kaiser et al. 2017. Learning to Remember Rare Events. ICLR

  • Note:

Use 𝑀! and 𝛾 value less than 1 to prevent losing the cumulative information. Previous work using 𝑀" reported indifference with using a uniform prior.

Cohn-Gordon et al. 2018. Pragmatically Informative Image Captioning With Character-Level Inference. NAACL-HLT

slide-28
SLIDE 28

Self-Conscious Speaker 𝑻𝟐

  • 𝑈ℎ𝑓 𝒕𝒇𝒎𝒈 𝒅𝒑𝒐𝒕𝒅𝒋𝒑𝒗𝒕 𝑡𝑞𝑓𝑏𝑙𝑓𝑠

𝑇&

$ 𝑣$

𝑗, ℎ, 𝑣%$) ∝ 𝑀#

$ 𝑗 ℎ, 𝑣'$, 𝑞$)( * 𝑇# $ 𝑣$ 𝑗, ℎ, 𝑣%$)

Base Speaker: 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

Imaginary Listener: 𝑀!

" (𝑗|𝑣$", ℎ, 𝑞")

Self-Conscious Speaker: 𝑇%

"

Persona: 𝑗 Dialogue History: ℎ

Learned Distractor Personas: 𝑗′

Speaker’s Utterance: 𝑣" ∝ 𝑀!

" 𝑗 ℎ, 𝑣$", 𝑞" &

∗ 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")

𝑞!"#(𝑗)

The posterior distribution

Intensity of Self-consciousness = Controlling the amount of the listener’s information

slide-29
SLIDE 29

Experiments: Dialogue NLI Evaluation Set PersonaChat Human Evaluation

Zhang et al. 2018. Personalizing Dialogue Agents: I have a dog, do you have pets too? ACL Welleck et al. 2019. Dialogue Natural Language Inference. ACL

slide-30
SLIDE 30

Results on Dialogue NLI

𝑇-: Base speaker model: Lost In Conversation & Transfer Transfo 𝑇.: Self-conscious speaker +DM: Distractor Memory

Task: 31 candidate utterances given. (1 ground-truth, 10 entailing, 10 neutral, 10 contradicting utterance) The model selects the best utterance by perplexity The proportion of selecting Ground-truth (Hits@1) Entailing utterance (Entail@1) Contradicting utterance (Contradict@1)

Alexander Tselousov and Sergey Golovanov. 2019. Lost In Conversation. Wolf et al. 2019. TransferTransfo: A Transfer Learning Approach for Neural Network Based Conversational Agents. arXiv

slide-31
SLIDE 31

Results on PersonaChat

𝑇-: Base speaker model: Lost In Conversation & Transfer Transfo 𝑇.: Self-conscious speaker +DM: Distractor Memory

C: consistency score, evaluation with pretrained NLI model

Madotto et al. 2019. Personalizing Dialogue Agents via Meta-Learning. ACL

slide-32
SLIDE 32

Results on Human Evaluation

Consistency: Is the response consistent? Engagingness: How much do you like the response?

  • n TransferTransfo model

Numbers in parentheses are standard error We also report Bayesian calibrated scores to remove evaluator bias

Kulikov et al. 2019. Importance of Search and Evaluation Strategies in Neural Dialogue Modeling. INLG

slide-33
SLIDE 33

Controlling the Self-conscious agent: 𝛽 and 𝛾

slide-34
SLIDE 34

𝛽 controls the degree of copying the given condition text (=persona)

Appropriate value allows the condition text to blend smoothly in the generation

  • 𝑈ℎ𝑓 𝒕𝒇𝒎𝒈 𝒅𝒑𝒐𝒕𝒅𝒋𝒑𝒗𝒕 𝑡𝑞𝑓𝑏𝑙𝑓𝑠

𝑇&

$ 𝑣$

𝑗, ℎ, 𝑣%$) ∝ 𝑀#

$ 𝑗 ℎ, 𝑣'$, 𝑞$)( * 𝑇# $ 𝑣$ 𝑗, ℎ, 𝑣%$)

slide-35
SLIDE 35

0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0

β

6 8 10 12 14 16

Hits@1 (%)

LostInConv β Ablation

L0 (Ours) L1 Uniform S0

0.0 0.5 1.0 1.5 2.0 2.5 3.0 3.5 4.0

β

8 10 12 14 16 18 20

Hits@1 (%)

L0 (Ours) L1 Uniform S0

𝛾 andWorld prior 𝑞3(𝑗)

TransferTransfo LostInConv Value equal to 1 or slightly less updating the world prior with 𝑀# is appropriate for incremental decoding

  • 𝐵𝑜 𝑗𝑛𝑏𝑕𝑗𝑜𝑏𝑠𝑧 𝑚𝑗𝑡𝑢𝑓𝑜𝑓𝑠

𝑀!

" 𝑗 ℎ, 𝑣$", 𝑞") ∝

𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")* J 𝑞"(𝑗)

∑+'∈- 𝑇!

" 𝑣" 𝑗, ℎ, 𝑣#")* J 𝑞" (𝑗.)

slide-36
SLIDE 36

Concluding Remarks

  • Introduced an unsupervised method for improving consistency inspired by

social cognition and pragmatics

→ Requiring no additional annotations nor external models

  • Further extended the Rational Speech Acts framework

→ Learning to provide distractors and different update for world prior

  • Extensive experiments on Dialogue NLI, PersonaChat and Human Evalution

→ Significantly reduced contradiction and improved ground-truth accuracy

slide-37
SLIDE 37

Thank you

Images and Icons are from Facebook ConvAI2 challenge, Nhor Phai, and Vincent Le Moign.