SECURITY AND PRIVACY OF MACHINE LEARNING Ian Goodfellow Staff - - PowerPoint PPT Presentation

security and privacy of machine learning
SMART_READER_LITE
LIVE PREVIEW

SECURITY AND PRIVACY OF MACHINE LEARNING Ian Goodfellow Staff - - PowerPoint PPT Presentation

#RSAC SESSION ID: SECURITY AND PRIVACY OF MACHINE LEARNING Ian Goodfellow Staff Research Scientist Google Brain @goodfellow_ian Machine Learning and Security #RSAC Machine Learning for Security Security against Machine Learning h 1 h 1 y


slide-1
SLIDE 1 SESSION ID: #RSAC

Ian Goodfellow

SECURITY AND PRIVACY OF MACHINE LEARNING

Staff Research Scientist Google Brain @goodfellow_ian

slide-2
SLIDE 2 (Goodfellow 2018) #RSAC

Machine Learning and Security

2

y y h1 h1 x1 x1 h2 h2 x2 x2

Machine Learning for Security Malware detection Intrusion detection … Security against Machine Learning

y y h1 h1 x1 x1 h2 h2 x2 x2

Password guessing Fake reviews …

slide-3
SLIDE 3 (Goodfellow 2018) #RSAC

Security of Machine Learning

3

y y h1 h1 x1 x1 h2 h2 x2 x2
slide-4
SLIDE 4 (Goodfellow 2018) #RSAC

An overview of a field

4

This presentation summarizes the work of many people, not just my own / my collaborators Download the slides for this link to extensive references The presentation focuses on the concepts, not the history or the inventors

slide-5
SLIDE 5 (Goodfellow 2018) #RSAC

Machine Learning Pipeline

5

X θ

x ˆ y

Training data Learning algorithm Learned parameters Test input Test output

slide-6
SLIDE 6 (Goodfellow 2018) #RSAC

Privacy of Training Data

6

X

θ ˆ X

slide-7
SLIDE 7 (Goodfellow 2018) #RSAC

Defining (ε, δ)-Differential Privacy

7

(Abadi 2017)

slide-8
SLIDE 8 (Goodfellow 2018) #RSAC

Private Aggregation of Teacher Ensembles

8

(Papernot et al 2016)

slide-9
SLIDE 9 (Goodfellow 2018) #RSAC

Training Set Poisoning

9

x X

θ ˆ y

slide-10
SLIDE 10 (Goodfellow 2018) #RSAC

ImageNet Poisoning

10

(Koh and Liang 2017)

slide-11
SLIDE 11 (Goodfellow 2018) #RSAC

Adversarial Examples

11

X θ

x

ˆ y

slide-12
SLIDE 12 (Goodfellow 2018) #RSAC

Model Theft

12

X θ x ˆ

y

ˆ θ

slide-13
SLIDE 13 (Goodfellow 2018) #RSAC

Model Theft++

13

X θ x ˆ y ˆ θ ˆ X

x

slide-14
SLIDE 14 (Goodfellow 2018) #RSAC

Deep Dive on Adversarial Examples

14

...solving CAPTCHAS and reading addresses... ...recognizing objects and faces…. (Szegedy et al, 2014) (Goodfellow et al, 2013) (Taigmen et al, 2013) (Goodfellow et al, 2013)

and other tasks... Since 2013, deep neural networks have matched human performance at...

slide-15
SLIDE 15 (Goodfellow 2018) #RSAC

Adversarial Examples

15

slide-16
SLIDE 16 (Goodfellow 2018) #RSAC

Turning objects into airplanes

16

slide-17
SLIDE 17 (Goodfellow 2018) #RSAC

Attacking a linear model

17

slide-18
SLIDE 18 (Goodfellow 2018) #RSAC

Wrong almost everywhere

18

slide-19
SLIDE 19 (Goodfellow 2018) #RSAC

Cross-model, cross-dataset transfer

19

slide-20
SLIDE 20 (Goodfellow 2018) #RSAC

Transfer across learning algorithms

20

(Papernot 2016)

slide-21
SLIDE 21 (Goodfellow 2018) #RSAC

Transfer attack

21

Train your

  • wn model

Target model with unknown weights, machine learning algorithm, training set; maybe non- differentiable Substitute model mimicking target model with known, differentiable function Adversarial examples Adversarial crafting against substitute Deploy adversarial examples against the target; transferability property results in them succeeding

slide-22
SLIDE 22 (Goodfellow 2018) #RSAC

Enhancing Transfer with Ensembles

22

(Liu et al, 2016)

slide-23
SLIDE 23 (Goodfellow 2018) #RSAC

Transfer to the Human Brain

23

(Elsayed et al, 2018)

slide-24
SLIDE 24 (Goodfellow 2018) #RSAC

Transfer to the Physical World

24

(Kurakin et al, 2016)

slide-25
SLIDE 25 (Goodfellow 2018) #RSAC

Adversarial Training

25 50 100 150 200 250 300 Training time (epochs) 10−2 10−1 100 Test misclassification rate

Train=Clean, Test=Clean Train=Clean, Test=Adv Train=Adv, Test=Clean Train=Adv, Test=Adv

slide-26
SLIDE 26 (Goodfellow 2018) #RSAC

Adversarial Training vs Certified Defenses

26

Adversarial Training:

Train on adversarial examples This minimizes a lower bound on the true worst-case error Achieves a high amount of (empirically tested) robustness on small to medium datasets

Certified defenses

Minimize an upper bound on true worst-case error Robustness is guaranteed, but amount of robustness is small Verification of models that weren’t trained to be easy to verify is hard

slide-27
SLIDE 27 (Goodfellow 2018) #RSAC

Limitations of defenses

27

Even certified defenses so far assume unrealistic threat model

Typical model: attacker can change input within some norm ball

Real attacks will be stranger, hard to characterize ahead of time

(Brown et al., 2017)

slide-28
SLIDE 28 (Goodfellow 2018) #RSAC

Clever Hans

28

(“Clever Hans, Clever Algorithms,” Bob Sturm)

slide-29
SLIDE 29 (Goodfellow 2018) #RSAC

Get involved!

29

https://github.com/tensorflow/cleverhans

slide-30
SLIDE 30 (Goodfellow 2018) #RSAC

Apply What You Have Learned

30

Publishing an ML model or a prediction API?

Is the training data sensitive? -> train with differential privacy

Consider how an attacker could cause damage by fooling your model

Current defenses are not practical Rely on situations with no incentive to cause harm / limited amount of potential harm