DATA ANALYTICS USING DEEP LEARNING GT 8803 // FALL 2019 // JOY - - PowerPoint PPT Presentation

data analytics using deep learning
SMART_READER_LITE
LIVE PREVIEW

DATA ANALYTICS USING DEEP LEARNING GT 8803 // FALL 2019 // JOY - - PowerPoint PPT Presentation

DATA ANALYTICS USING DEEP LEARNING GT 8803 // FALL 2019 // JOY ARULRAJ L E C T U R E # 2 0 : A D V E R S A R I A L T R A I N I N G administrivia Reminders Best project prize Quiz cancelled Guest lecture GT 8803 // Fall 2019


slide-1
SLIDE 1

DATA ANALYTICS USING DEEP LEARNING

GT 8803 // FALL 2019 // JOY ARULRAJ

L E C T U R E # 2 0 : A D V E R S A R I A L T R A I N I N G

slide-2
SLIDE 2

GT 8803 // Fall 2019

administrivia

  • Reminders

– Best project prize – Quiz cancelled – Guest lecture

2

slide-3
SLIDE 3

GT 8803 // Fall 2019

CREDITS

  • Slides based on a lecture by:

– Ian Goodfellow @ Google Brain

3

slide-4
SLIDE 4

GT 8803 // Fall 2019

OVERVIEW

  • What are adversarial examples?
  • Why do they happen?
  • How can they be used to compromise machine

learning systems?

  • What are the defenses?
  • How to use adversarial examples to improve

machine learning (even without adversary)?

4

slide-5
SLIDE 5

GT 8803 // Fall 2019

ADVERSARIAL EXAMPLES

5

slide-6
SLIDE 6

GT 8803 // Fall 2019 6 ...solving CAPTCHAS and reading addresses...

...recognizing

  • bjects

and faces….

(Szegedy et al, 2014) (Goodfellow et al, 2013) (Taigmen et al, 2013) (Goodfellow et al, 2013)

and other tasks... Since 2013, deep neural networks have matched human performance at...

slide-7
SLIDE 7

GT 8803 // Fall 2019

Adversarial Examples

7

Timeline: “Adversarial Classification” Dalvi et al 2004: fool spam filter “Evasion Attacks Against Machine Learning at Test Time” Biggio 2013: fool neural nets Szegedy et al 2013: fool ImageNet classifiers imperceptibly Goodfellow et al 2014: cheap, closed form attack

slide-8
SLIDE 8

GT 8803 // Fall 2019

Turning Objects into “Airplanes”

8

slide-9
SLIDE 9

GT 8803 // Fall 2019

Attacking a Linear Model

9

slide-10
SLIDE 10

GT 8803 // Fall 2019

Not just for neural nets

  • Linear models

– Logistic regression – Softmax regression – SVMs

  • Decision trees
  • Nearest neighbors

10

slide-11
SLIDE 11

GT 8803 // Fall 2019

Adversarial Examples from Overfitting

11

x x x

O O O

x

O

slide-12
SLIDE 12

GT 8803 // Fall 2019

Adversarial Examples from Overfitting

12

x x x

O O O O O

x

slide-13
SLIDE 13

GT 8803 // Fall 2019

Modern deep nets are very piecewise linear

13

Rectified linear unit Carefully tuned sigmoid Maxout LSTM

slide-14
SLIDE 14

GT 8803 // Fall 2019

Nearly Linear Responses in Practice

14

slide-15
SLIDE 15

GT 8803 // Fall 2019

Small inter-class distances

15

Clean example Perturbation Corrupted example All three perturbations have L2 norm 3.96 This is actually small. We typically use 7! Perturbation changes the true class Random perturbation does not change the class Perturbation changes the input to “rubbish class”

slide-16
SLIDE 16

GT 8803 // Fall 2019

The Fast Gradient Sign Method

16

slide-17
SLIDE 17

GT 8803 // Fall 2019

Maps of Adversarial and Random Cross-Sections

17

(collaboration with David Warde-Farley and Nicolas Papernot)

slide-18
SLIDE 18

GT 8803 // Fall 2019

18

Maps of Adversarial Cross-Sections

slide-19
SLIDE 19

GT 8803 // Fall 2019

19

Maps of RANDOM Cross-Sections

(collaboration with David Warde-Farley and Nicolas Papernot)

slide-20
SLIDE 20

GT 8803 // Fall 2019

Estimating the Subspace Dimensionality

20

slide-21
SLIDE 21

GT 8803 // Fall 2019

Clever Hans

21

(“Clever Hans, Clever Algorithms,” Bob Sturm)

slide-22
SLIDE 22

GT 8803 // Fall 2019

Wrong almost everywhere

22

slide-23
SLIDE 23

GT 8803 // Fall 2019

Adversarial Examples for RL

23

(Huang et al., 2017)

slide-24
SLIDE 24

GT 8803 // Fall 2019

High-Dimensional Linear Models

24

Weights Signs of weights Clean examples Adversarial examples

slide-25
SLIDE 25

GT 8803 // Fall 2019

Linear Models of ImageNet

25

(Andrej Karpathy, “Breaking Linear Classifiers on ImageNet”)

slide-26
SLIDE 26

GT 8803 // Fall 2019

RBFs behave more intuitively

26

slide-27
SLIDE 27

GT 8803 // Fall 2019

Cross-model, cross-dataset generalization

27

slide-28
SLIDE 28

GT 8803 // Fall 2019

Cross-technique transferability

28

(Papernot 2016)

slide-29
SLIDE 29

GT 8803 // Fall 2019

Transferability Attack

29

Train your own model Target model with unknown weights, machine learning algorithm, training set; maybe non-differentiable Substitute model mimicking target model with known, differentiable function Adversarial examples Adversarial crafting against substitute Deploy adversarial examples against the target; transferability property results in them succeeding

slide-30
SLIDE 30

GT 8803 // Fall 2019 30

(Papernot 2016)

slide-31
SLIDE 31

GT 8803 // Fall 2019

Enhancing Transfer With Ensembles

31

(Liu et al, 2016)

slide-32
SLIDE 32

GT 8803 // Fall 2019

Adversarial Examples in the Human Brain

32

(Pinna and Gregory, 2002)

These are concentric circles, not intertwined spirals.

slide-33
SLIDE 33

GT 8803 // Fall 2019

Practical Attacks

  • Fool real classifiers trained by remotely

hosted API (MetaMind, Amazon, Google)

  • Fool malware detector networks
  • Display adversarial examples in the physical

world and fool machine learning systems that perceive them through a camera

33

slide-34
SLIDE 34

GT 8803 // Fall 2019

Adversarial Examples in the Physical World

34

slide-35
SLIDE 35

GT 8803 // Fall 2019

Failed defenses

35

Weight decay Adding noise at test time Adding noise at train time Dropout Ensembles Multiple glimpses Generative pretraining Removing perturbation with an autoencoder Error correcting codes Confidence-reducing perturbation at test time Various non-linear units Double backprop

slide-36
SLIDE 36

GT 8803 // Fall 2019

Generative Modeling is not SufficienT

36

slide-37
SLIDE 37

GT 8803 // Fall 2019

Universal approximator theorem

37

Neural nets can represent either function: Maximum likelihood doesn’t cause them to learn the right function. But we can fix that...

slide-38
SLIDE 38

GT 8803 // Fall 2019

ADVERSARIAL TRAINING

38

slide-39
SLIDE 39

GT 8803 // Fall 2019

Training on Adversarial Examples

39

slide-40
SLIDE 40

GT 8803 // Fall 2019

Adversarial Training of other Models

  • Linear models: SVM / linear regression cannot learn a

step function, so adversarial training is less useful, very similar to weight decay

  • k-NN: adversarial training is prone to overfitting.
  • Takeway: neural nets can actually become more secure

than other models. Adversarially trained neural nets have the best empirical success rate on adversarial examples of any machine learning model.

40

slide-41
SLIDE 41

GT 8803 // Fall 2019

Weaknesses Persist

41

slide-42
SLIDE 42

GT 8803 // Fall 2019

Adversarial Training

42

Labeled as bird Decrease probability

  • f bird class

Still has same label (bird)

slide-43
SLIDE 43

GT 8803 // Fall 2019

Virtual Adversarial Training

43

Unlabeled; model guesses it’s probably a bird, maybe a plane Adversarial perturbation intended to change the guess New guess should match old guess (probably bird, maybe plane)

slide-44
SLIDE 44

GT 8803 // Fall 2019

Text Classification with VAT

44

7.70 7.20 7.40 7.12 7.05 6.97 6.68

6.00 6.50 7.00 7.50 8.00

Earlier SOTA Our baseline Virtual Adversarial Both + bidirectional model

RCV1 Misclassification Rate

Zoomed in for legibility

slide-45
SLIDE 45

GT 8803 // Fall 2019 45

Universal engineering machine (model-based optimization)

Training data Extrapolation Make new inventions by finding input that maximizes model’s predicted performance

slide-46
SLIDE 46

GT 8803 // Fall 2019 46

cleverhans

Open-source library available at: https://github.com/openai/cleverhans Built on top of TensorFlow (Theano support anticipated) Standard implementation of attacks, for adversarial training and reproducible benchmarks

slide-47
SLIDE 47

GT 8803 // Fall 2019

Conclusion

  • Attacking is easy
  • Defending is difficult
  • Adversarial training provides regularization and

semi-supervised learning

  • The out-of-domain input problem is a bottleneck for

model-based optimization generally

47