Sicherheitslcken in der knstlichen Intelligenz Konrad Rieck, TU - - PowerPoint PPT Presentation

sicherheitsl cken in der k nstlichen intelligenz
SMART_READER_LITE
LIVE PREVIEW

Sicherheitslcken in der knstlichen Intelligenz Konrad Rieck, TU - - PowerPoint PPT Presentation

Sicherheitslcken in der knstlichen Intelligenz Konrad Rieck, TU Braunschweig Keynote 1oth German OWASP Day 2018 The AI Hype Hype around artificial intelligence and deep learning Amazing progress of


slide-1
SLIDE 1

Konrad Rieck, TU Braunschweig

Keynote — 1oth German OWASP Day 2018

Sicherheitslücken in der künstlichen Intelligenz

slide-2
SLIDE 2

Page

The AI Hype

  • Hype around artificial intelligence and deep learning
  • Amazing progress of machine learning techniques
  • Novel learning concepts, strategies and algorithms
  • Impressive results in computer vision and linguistics



 
 
 
 
 


2

Autonomous cars 
 and drones Medical diagnosis 
 and prediction Virtual Assistants
 (Siri, Alexa & Friends) Cool stuff! 
 But is this secure?

slide-3
SLIDE 3

Page

Overview

  • What we will cover in this talk ...
  • Brief introduction to machine learning

How do computers learn something?

  • Attacks against machine learning

How do I break machine learning?

  • Current defenses for machine learning

Is there anything we can do?

3

slide-4
SLIDE 4

Page

Machine Learning

A Brief Introduction

4

slide-5
SLIDE 5

Page

AI and Machine Learning

  • Machine learning = branch of artificial intelligence
  • Computer science intersecting with statistics
  • No science fiction and no black magic, please! 



 
 
 
 
 


5

WOPR HAL 9000 T-800

ML

AI

slide-6
SLIDE 6

Page

How do computers learn?

  • An example: Handwriting recognition



 
 
 


  • Automatic inference of dependencies from data
  • Generalization of dependencies; ↯ not simple memorization
  • Dependencies represented by learning model
  • Application of learning model to unseen data

6

LeJers WriJen 
 shapes

slide-7
SLIDE 7

Page

Learning as a Process

  • Overview of learning process
  • Learning: Inference of model Θ from data X and labels Y
  • Application: Model Θ parametrizes prediction function fΘ: X → Y

7

Learning Data; Labels

X × Y Θ

Application Predictions Novel Data

X fΘ(X)

8

Train Apply

slide-8
SLIDE 8

Page

Classification

  • Classification = categorization of objects into classes
  • Most popular form of learning in practical applications
  • Large diversity of concepts, models and algorithms
  • Geometric interpretation
  • Feature space X = ℝN
  • Labels Y = {-1, +1}
  • Feature space partitioned


by prediction function f

8

  • 1

+1

slide-9
SLIDE 9

Page

Different Learning Models

9

Quadratic functions

Neural networks Decision trees

fΘ fΘ

slide-10
SLIDE 10

Page

Attacks against Machine Learning

Let’s break things ...

10

slide-11
SLIDE 11

Page

Security and Machine Learning

  • Originally no notion of security in machine learning
  • Learning algorithms designed for peaceful environments
  • Optimization of average-case errors; ↯ not worst-case errors
  • New research direction: Adversarial machine learning
  • Attacks and defenses for learning algorithms
  • History of ~10 years (good overview by Biggio & Roli)
  • Recent hype around deep learning and adversarial examples

11 (Biggio & Roli, PR’18)

slide-12
SLIDE 12

Page

Vulnerabilities and Attacks

  • Different types of vulnerabilities
  • Attacks possible during learning and application phase



 
 
 
 
 
 
 


12

Learning Data; Labels

X × Y Θ

Application Predictions Novel Data

X fΘ(X)

8

3 1 2

slide-13
SLIDE 13

Page

Attack: Adversarial Examples

  • Attacks misleading the prediction function
  • Minimal perturbation t of input x inducing misclassification


  • Attacks effective and robust
  • Small perturbations sufficient
  • Many learning algorithms vulnerable
  • Attacks against integrity of prediction

13

1

arg min

t

d(t) s.t. fΘ(x + t) = y*

fΘ x x + t

(Szegedy et al.,’14)

slide-14
SLIDE 14

Page

  • Adversarial examples generated using trivial algorithm
  • Greedy search for decision boundary by changing pixels
  • Two variants: sparse and dense (constrained) changes



 
 
 
 
 
 


A Toy Example

14

Sparse attack against SVM Dense attack against SVM

1

slide-15
SLIDE 15

Page

A Semi-Toy Example

  • Adversarial examples for object recognition
  • State-of-the-art attack against deep neural network
  • Perturbations visible but irrelevant to human observer



 
 
 
 
 


15

Detected: Airplane Detected: Car Detected: Truck Detected: Dog

1

slide-16
SLIDE 16

Page

A Realistic Example

  • Attack against state-of-the-art face recognition
  • Perturbations constrained to surface of eyeglasses
  • Surprising impersonation attacks possible



 
 
 
 
 
 


16

1

Detected:
 Milla Jovovich Detected:
 Milla Jovovich

(Sharif et al., CCS’16)

slide-17
SLIDE 17

Page

Attack: Model Stealing

  • Attacks “stealing” the learning model
  • Reconstruction of model using small set of inputs Z



 


  • Further related attacks
  • Membership and property inference
  • Model inversion attacks
  • Attacks against confidentiality of model

17

2

arg min

Z |Z|

s.t. Θ ≈ r(Z, fΘ)

Z

(Tramer et al., USENIX Security’16)

slide-18
SLIDE 18

Page

A Toy Example

  • Model stealing against linear classifiers
  • Exploration of prediction function with orthogonal inputs
  • Least squares approximation of prediction function



 
 
 
 
 


18

2

Model of 
 linear SVM Reconstructed model

slide-19
SLIDE 19

Page

A Realistic Example

19

2

(Fredrikson et al., CCS’15)

Image in
 training set Reconstructed image

  • Model inversion attack against face recognition
  • Attack reconstructs matching input data for prediction
  • Not perfect but still scary — 80% extracted faces recognized 



 
 
 
 
 
 


slide-20
SLIDE 20

Page

Attack: Poisoning and Backdoors

  • Attacks manipulating the learning model
  • Manipulation using small set of “poisoned” training data Z


  • Attack only possible if ...
  • Training data or model accessible

→ Supply chain of learning technology

  • Attacks against integrity of model

20

(Biggio et al., ICML’12)

arg min

Z |Z|

s.t. Θ* = g(X ∪ Z, Y)

3

slide-21
SLIDE 21

Page

A Toy Example

  • Poisoning of a linear classifier with trivial algorithm
  • Simple backdoor example added to training dataset
  • Poisoning of dataset increased until backdoor triggered



 
 
 
 
 


21

Poisoned 
 model Backdoor pattern (= 8)

3

slide-22
SLIDE 22

Page

A Semi-Toy Example

22

Trigger

(Liu et al., NDSS’18)

Backdoored navigation

  • Poisoning of decision system in a driving simulation
  • Decision system trained to navigate based on environment
  • Artificial traffic sign triggers strong steering to right

3

slide-23
SLIDE 23

Page

A Realistic Example

  • Poisoning of traffic-sign recognition
  • State-of-the-art backdoor for deep neural networks
  • Backdoor implanted through retraining with poisoned data

23

Misclassified stop sign Very small trigger

3

(Gu et al., MLSEC’17)

slide-24
SLIDE 24

Page

Defenses for Machine Learning

Let’s try to fix this ...

24

slide-25
SLIDE 25

Page

Defenses

  • Defense is a tough problem
  • Input data to system under control of adversary
  • Even training data hard to verify and sanitize
  • Often direct access to prediction function
  • Two defense strategies
  • Integrated defenses = Attack-resilient learning algorithms
  • Operational defenses = Security-aware application of learning
  • No strong defenses currently known!

25

slide-26
SLIDE 26

Page

Complexity and randomization

  • Defense: Complexity
  • Prediction function obfuscated
  • Addition of complexity (e.g. fractals)
  • Obfuscation of gradients
  • Defense: Randomization
  • Prediction function randomized
  • Noise added to output
  • Random feature selection

26

  • Both defenses ineffective
  • Approximation of 


true prediction function

(Athalye et al., ICML’18)

slide-27
SLIDE 27

Page

Stateful Application

  • Defense: Stateful Application
  • Access to function monitored
  • Input data associated with users
  • Detection of unusual behavior
  • Limited applicability in practice
  • Only feasible with remote access to learning
  • Concept for authentication and identify binding necessary
  • Sybial attacks (multiple accounts) still a problem

27

User 1

slide-28
SLIDE 28

Page

Security-Aware Testing

  • Defense: Better testing for models
  • Testing around boundary
  • Testing of corner cases
  • Analysis of neural coverage
  • Defense: Differential testing
  • Training of multiple models
  • Analysis of differences between learned models
  • But: Inherent limitations of testing approaches

28

slide-29
SLIDE 29

Page

Conclusions

29

slide-30
SLIDE 30

Page

Conclusions

  • Take-Away: Machine learning is insecure!
  • Learning algorithms not smart — despite the hype
  • Learned models ≠ human perception and understanding
  • Integrity and confidentiality not guaranteed
  • Take-Away: Security research urgently needed!
  • Current defenses still largely ineffective
  • Demand for better integrated and operational security
  • Testing and verification of learning promising direction

30

slide-31
SLIDE 31

Page

Thanks! Questions?

31

slide-32
SLIDE 32

Page

References

  • Battista Biggio, Fabio Roli. Wild Patterns: Ten Years After the Rise of Adversarial Machine
  • Learning. Pattern Recognition, 2018
  • Szegedy et al. Intriguing properties of neural networks. ArXiv 2014
  • Sharif et al. Accessorize to a Crime: Real and Stealthy Attacks on State-of-the-Art Face
  • Recognition. ACM CCS 2016.
  • Tramer et al. Stealing Machine Learning Models via Prediction APIs. USENIX Security 2016
  • Fredriskon et al. Model Inversion Attacks that Exploit Confidence Information and Basic
  • Countermeasures. ACM CCS 2015.
  • Biggio et al. Poisoning Attacks against Support Vector Machines. ICML 2012
  • Liu et al. Trojaning Attack on Neural Networks. NDSS 2018.
  • Gu et al. BadNets: Identifying Vulnerabilities in the Machine Learning Model Supply
  • Chain. Machine Learning and Security Workshop 2017.
  • Athalye et al. Obfuscated Gradients Give a False Sense of Security: Circumventing

Defenses to Adversarial Examples. ICML 2018.

32