provably robust boosted decision stumps and trees against
play

Provably Robust Boosted Decision Stumps and Trees against - PowerPoint PPT Presentation

Provably Robust Boosted Decision Stumps and Trees against Adversarial Attacks Maksym Andriushchenko (EPFL ) Matthias Hein (University of T ubingen) Work done at the University of T ubingen SMLD 2019, NeurIPS 2019 Stumps Trees


  1. Provably Robust Boosted Decision Stumps and Trees against Adversarial Attacks Maksym Andriushchenko (EPFL ∗ ) Matthias Hein (University of T¨ ubingen) ∗ Work done at the University of T¨ ubingen SMLD 2019, NeurIPS 2019 Stumps Trees Plain boosted stumps Robust boosted stumps Plain boosted trees Robust boosted trees 1.00 1.00 1.00 1.00 0.75 0.75 0.75 0.75 0.50 0.50 0.50 0.50 0.25 0.25 0.25 0.25 0.00 0.00 0.00 0.00 0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00 Maksym Andriushchenko (EPFL) 1

  2. Adversarial vulnerability Source: Goodfellow et al, “Explaining and Harnessing Adversarial Examples”, 2014 Maksym Andriushchenko (EPFL) 2

  3. Adversarial vulnerability Source: Goodfellow et al, “Explaining and Harnessing Adversarial Examples”, 2014 Problem : small changes in the input ⇒ large changes in the output Maksym Andriushchenko (EPFL) 2

  4. Adversarial vulnerability Source: Goodfellow et al, “Explaining and Harnessing Adversarial Examples”, 2014 Problem : small changes in the input ⇒ large changes in the output Topic of active research for neural networks and image recognition, but what about other domains and other classifiers ? Maksym Andriushchenko (EPFL) 2

  5. Motivation: other domains (going beyond images) Some input feature values can be incorrect : measurement noise, a human mistake, an adversarially crafted change, etc. Maksym Andriushchenko (EPFL) 3

  6. Motivation: other domains (going beyond images) Some input feature values can be incorrect : measurement noise, a human mistake, an adversarially crafted change, etc. For high-stakes decision making, it’s necessary to ensure a reasonable worst-case error rate under possible noise perturbations Maksym Andriushchenko (EPFL) 3

  7. Motivation: other domains (going beyond images) Some input feature values can be incorrect : measurement noise, a human mistake, an adversarially crafted change, etc. For high-stakes decision making, it’s necessary to ensure a reasonable worst-case error rate under possible noise perturbations The expected perturbation range can be specified by domain experts Maksym Andriushchenko (EPFL) 3

  8. Motivation: other classifiers Our paper : we concentrate on boosted decision stumps and trees Maksym Andriushchenko (EPFL) 4

  9. Motivation: other classifiers Our paper : we concentrate on boosted decision stumps and trees They are widely adopted in practice – implementations like XGBoost or LightGBM are used almost in every Kaggle competition Maksym Andriushchenko (EPFL) 4

  10. Motivation: other classifiers Our paper : we concentrate on boosted decision stumps and trees They are widely adopted in practice – implementations like XGBoost or LightGBM are used almost in every Kaggle competition Moreover, boosted trees are interpretable which is also an important practical aspect. Who wants to deploy a black-box? Maksym Andriushchenko (EPFL) 4

  11. Motivation: other classifiers Our paper : we concentrate on boosted decision stumps and trees They are widely adopted in practice – implementations like XGBoost or LightGBM are used almost in every Kaggle competition Moreover, boosted trees are interpretable which is also an important practical aspect. Who wants to deploy a black-box? = ⇒ it is important to develop boosted trees which are robust , but first we need to understand the reason of their vulnerability Maksym Andriushchenko (EPFL) 4

  12. Motivation: other classifiers Our paper : we concentrate on boosted decision stumps and trees They are widely adopted in practice – implementations like XGBoost or LightGBM are used almost in every Kaggle competition Moreover, boosted trees are interpretable which is also an important practical aspect. Who wants to deploy a black-box? = ⇒ it is important to develop boosted trees which are robust , but first we need to understand the reason of their vulnerability So why do adversarial examples exist? Maksym Andriushchenko (EPFL) 4

  13. Plain boosted stumps Robust boosted stumps 1.00 1.00 0.75 0.75 0.50 0.50 0.25 0.25 0.00 0.00 0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00 Understanding adversarial vulnerability What goes wrong and how to fix it? Maksym Andriushchenko (EPFL) 5

  14. Understanding adversarial vulnerability What goes wrong and how to fix it? We would like to have a large geometric margin for every point Plain boosted stumps Robust boosted stumps 1.00 1.00 0.75 0.75 0.50 0.50 0.25 0.25 0.00 0.00 0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00 Maksym Andriushchenko (EPFL) 5

  15. Understanding adversarial vulnerability What goes wrong and how to fix it? We would like to have a large geometric margin for every point Plain boosted stumps Robust boosted stumps 1.00 1.00 0.75 0.75 0.50 0.50 0.25 0.25 0.00 0.00 0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00 Empirical risk minimization does not distinguish the two types of solutions ⇒ we need to use a robust objective Maksym Andriushchenko (EPFL) 5

  16. Understanding adversarial vulnerability What goes wrong and how to fix it? We would like to have a large geometric margin for every point Plain boosted stumps Robust boosted stumps 1.00 1.00 0.75 0.75 0.50 0.50 0.25 0.25 0.00 0.00 0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00 Empirical risk minimization does not distinguish the two types of solutions ⇒ we need to use a robust objective Let’s formalize the problem! Maksym Andriushchenko (EPFL) 5

  17. ✶ ✶ Adversarial robustness What is an adversarial example ? Consider x ∈ R d , y ∈ {− 1 , 1 } , classifier f : R d → R , some L p -norm threshold ǫ : δ ∈ R d yf ( x + δ ) min � δ � p ≤ ǫ, x + δ ∈ C Maksym Andriushchenko (EPFL) 6

  18. ✶ ✶ Adversarial robustness What is an adversarial example ? Consider x ∈ R d , y ∈ {− 1 , 1 } , classifier f : R d → R , some L p -norm threshold ǫ : δ ∈ R d yf ( x + δ ) min � δ � p ≤ ǫ, x + δ ∈ C Assume x is correctly classified ( yf ( x ) > 0), then x + δ ∗ is an adversarial example if x + δ ∗ is incorrectly classified ( yf ( x + δ ∗ ) < 0) Maksym Andriushchenko (EPFL) 6

  19. Adversarial robustness What is an adversarial example ? Consider x ∈ R d , y ∈ {− 1 , 1 } , classifier f : R d → R , some L p -norm threshold ǫ : δ ∈ R d yf ( x + δ ) min � δ � p ≤ ǫ, x + δ ∈ C Assume x is correctly classified ( yf ( x ) > 0), then x + δ ∗ is an adversarial example if x + δ ∗ is incorrectly classified ( yf ( x + δ ∗ ) < 0) How to measure robustness? Robust test error (RTE): n n 1 1 � � ✶ yf ( x ) < 0 → ✶ yf ( x + δ ∗ ) < 0 n n i =1 i =1 � �� � � �� � standard zero-one loss robust zero-one loss Maksym Andriushchenko (EPFL) 6

  20. Adversarial robustness What is an adversarial example ? Consider x ∈ R d , y ∈ {− 1 , 1 } , classifier f : R d → R , some L p -norm threshold ǫ : δ ∈ R d yf ( x + δ ) min � δ � p ≤ ǫ, x + δ ∈ C Assume x is correctly classified ( yf ( x ) > 0), then x + δ ∗ is an adversarial example if x + δ ∗ is incorrectly classified ( yf ( x + δ ∗ ) < 0) How to measure robustness? Robust test error (RTE): n n 1 1 � � ✶ yf ( x ) < 0 → ✶ yf ( x + δ ∗ ) < 0 n n i =1 i =1 � �� � � �� � standard zero-one loss robust zero-one loss Finding δ ∗ : non-convex opt. problem for NNs and BTs. Exact mixed integer formulations exist for ReLU-NNs and BTs ( slow ). Maksym Andriushchenko (EPFL) 6

  21. Training adversarially robust models Robust optimization problem wrt the set ∆( ǫ ): n � δ ∈ ∆( ǫ ) L ( f ( x i + δ ; θ ) , y i ) min max θ i =1 Maksym Andriushchenko (EPFL) 7

  22. Training adversarially robust models Robust optimization problem wrt the set ∆( ǫ ): n � δ ∈ ∆( ǫ ) L ( f ( x i + δ ; θ ) , y i ) min max θ i =1 L is a usual margin-based loss function (cross-entropy, exp. loss, etc) Maksym Andriushchenko (EPFL) 7

  23. Training adversarially robust models Robust optimization problem wrt the set ∆( ǫ ): n � δ ∈ ∆( ǫ ) L ( f ( x i + δ ; θ ) , y i ) min max θ i =1 L is a usual margin-based loss function (cross-entropy, exp. loss, etc) ǫ = 0 = ⇒ just well-known Empirical Risk Minimization Maksym Andriushchenko (EPFL) 7

  24. Training adversarially robust models Robust optimization problem wrt the set ∆( ǫ ): n � δ ∈ ∆( ǫ ) L ( f ( x i + δ ; θ ) , y i ) min max θ i =1 L is a usual margin-based loss function (cross-entropy, exp. loss, etc) ǫ = 0 = ⇒ just well-known Empirical Risk Minimization Goal : small loss ( ⇒ large margin) not only at x i , but for every x i + δ ∈ ∆( ǫ ) Maksym Andriushchenko (EPFL) 7

  25. Training adversarially robust models Robust optimization problem wrt the set ∆( ǫ ): n � δ ∈ ∆( ǫ ) L ( f ( x i + δ ; θ ) , y i ) min max θ i =1 L is a usual margin-based loss function (cross-entropy, exp. loss, etc) ǫ = 0 = ⇒ just well-known Empirical Risk Minimization Goal : small loss ( ⇒ large margin) not only at x i , but for every x i + δ ∈ ∆( ǫ ) Adversarial training : approximately solve the robust loss = ⇒ minimization of a lower bound on the objective Maksym Andriushchenko (EPFL) 7

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend