Deep Adversarial Learning for NLP
William Wang Sameer Singh
With contributions from Jiwei Li.
1
Slides: http://tiny.cc/adversarial
NLP William Wang Sameer Singh Slides: http://tiny.cc/adversarial - - PowerPoint PPT Presentation
Deep Adversarial Learning for NLP William Wang Sameer Singh Slides: http://tiny.cc/adversarial With contributions from Jiwei Li. 1 Agenda Introduction, Background, and GANs (William, 90 mins) Adversarial Examples and Rules (Sameer, 75
With contributions from Jiwei Li.
1
Slides: http://tiny.cc/adversarial
mins)
2
Slides: http://tiny.cc/adversarial
3
there were 20+ times more papers mentioning “adversarial”, comparing to 2016.
during this period.
more than 100 papers on adversarial learning (approximately 1/3 of all adv. learning papers in NLP).
4
5
is closely related to, but limited to the following fields of study:
6
7
8
CycleGAN (Zhu et al., 2017)
9
CycleGAN (Zhu et al., 2017)
10
GauGAN (Park et al., 2019)
11
12
13
14
15
The main classifier predicts a label y from a text x, the attacker tries to recover some private information z contained in x from the latent representation used by the main classifier.
16
17
Wu, Bamman, Russell (EMNLP 2017).
18
Wu, Bamman, Russell (EMNLP 2017).
19
20
the classifier trying to detect the fake sample forger trying to produce some counterfeit material Image: https://ishmaelbelghazi.github.io/ALI/
21
22
D(x): the probability that x came from the data rather than generator
Goodfellow, et al., “Generative adversarial networks,” in NIPS, 2014.
D G
23
Discriminator Generator
Goodfellow, et al., “Generative adversarial networks,” in NIPS, 2014.
24
Goodfellow, et al., “Generative adversarial networks,” in NIPS, 2014.
D G
s.t.
25
MLE
and discriminator networks in NLP, it could be worse.
26
2017).
27
28
2019)
29
with dense rewards).
30
2018; Chen and Cardie, 2018; Tran and Nguyen, 2018; Cao et al., 2018; Li et al., 2018b)
al., 2018; Zellers et al., 2018)
al., 2018b; Shi et al., 2018a; Bekoulis et al., 2018)
31
32
Idea: use a mixture of generators and a multi-class discriminator.
33
34
0% 10% 20% 30% 40% 50% XE BLEU-RL CIDEr-RL GAN AREL
Turing Test
Win Unsure
35
36
37
Idea: use adversarial learning to iteratively learn better negative examples.
38
39
40
41
How old are you ? I don’t know what you are talking about I’m 25.
A human evaluator/ judge
42
Jl3
How old are you ? I don’t know what you are talking about I’m 25.
43
Jl3
How old are you ? I don’t know what you are talking about I’m 25.
P= 90% human generated P= 10% human generated
44
Jl3
Adversarial Learning in Image Generation (Goodfellow et al., 2014)
45
Jl3 Jl4
Generative Model (G)
how are you ? I’m fine . EOS
Encoding Decoding
eos I’m fine .
46
Generative Model (G)
how are you ? I’m fine . EOS
Encoding Decoding
eos I’m fine .
Discriminative Model (D)
how are you ? eos I’m fine .
P= 90% human generated
47
Generative Model (G)
how are you ? I’m fine . EOS
Encoding Decoding
eos I’m fine .
Discriminative Model (D)
how are you ? eos I’m fine .
Reward P= 90% human generated
48
REINFORCE Algorithm (William,1992)
Generative Model (G)
how are you ? I’m fine EOS
Encoding Decoding
eos I’m fine .
49
Update the Discriminator Update the Generator
The discriminator forces the generator to produce correct responses
50
The previous RL model only perform better on multi-turn conversations
51
vs a vanilla generation model Adversarial Win Adversarial Lose Tie 62% 18% 20%
52
53
54
55
56
57
58
disciplinary research area, and it is highly related to many subareas in NLP.
there are both challenges and opportunities in GANs for NLP.
has obtained promising results.
advances of representation learning, reinforcement learning, and self-supervised learning techniques in NLP.
59
60
61