a closer look at adversarial examples for separated data
play

A Closer Look at Adversarial Examples for Separated Data Kamalika - PowerPoint PPT Presentation

A Closer Look at Adversarial Examples for Separated Data Kamalika Chaudhuri University of California, San Diego Adversarial Examples Gibbon Panda Small perturbation to legitimate inputs causing misclassification Adversarial Examples Can


  1. A Closer Look at Adversarial Examples for Separated Data Kamalika Chaudhuri University of California, San Diego

  2. Adversarial Examples Gibbon Panda Small perturbation to legitimate inputs causing misclassification

  3. Adversarial Examples Can potentially lead to serious safety issues

  4. Adversarial Examples: State of the Art A large number of attacks A few defenses Not much understanding on why adversarial examples arise

  5. Adversarial Examples: State of the Art A large number of attacks A few defenses Not much understanding on why adversarial examples arise This talk: A closer look

  6. Background: Classification Given: ( x i, y i ) Vector of Discrete features Labels Find: Prediction rule in a class to predict y from x

  7. Background: The Statistical Learning Framework - + Training and test data drawn from an underlying distribution D

  8. Background: The Statistical Learning Framework - + Training and test data drawn from an underlying distribution D Goal: Find classifier f to maximize accuracy ( x,y ) ∼ D ( f ( x ) = y ) Pr

  9. Measure of Robustness: L p norm A classifier f is robust with radius r at x if it predicts f(x) for all x’ in k x � x 0 k p  r

  10. Why do we have adversarial examples?

  11. Why do we have adversarial examples? - Data Distributional distribution Robustness +

  12. Why do we have adversarial examples? - Data Distributional distribution Robustness + Too few - Finite Sample samples Robustness +

  13. Why do we have adversarial examples? - Data Distributional distribution Robustness + Too few - Finite Sample samples Robustness + Algorithmic Bad - Robustness algorithm +

  14. Why do we have adversarial examples? - Data Distributional distribution Robustness + Are classes separated in real data?

  15. r-Separation 2r 2r Data distribution D is r-separated if for any (x, y) and (x’, y’) drawn from D y 6 = y 0 = ) k x � x 0 k � 2 r

  16. r-Separation 2r 2r 2r Data distribution D is r-separated if for any (x, y) and (x’, y’) drawn from D y 6 = y 0 = ) k x � x 0 k � 2 r r-separation means accurate and robust at radius r classifier possible!

  17. Real Data is r-Separated Dataset Separation Typical r 0.1 MNIST 0.74 0.03 CIFAR10 0.21 0.03 SVHN* 0.09 0.005 ResImgnet* 0.18 Separation = min distance between any two points in different classes

  18. Robustness for r-separated data: Two Settings Non-parametric Methods

  19. Non-Parametric Methods k-Nearest Neighbors Decision Trees Others: Random Forests, Kernel classifiers, etc

  20. What is known about Nonparametric Methods?

  21. The Bayes Optimal Classifier Classifier with maximum accuracy on data distribution Only reachable in the large sample limit

  22. What is known about Non-Parametrics? With growing training data, accuracy of non-parametric methods converge to accuracy of the Bayes Optimal

  23. What about Robustness? Prior work: Attacks and defenses for specific classifiers Our work: General conditions when we can get robustness

  24. What is the goal of robust classification?

  25. What is the goal of robust classification? Bayes optimal undefined outside distribution Bayes optimal

  26. The r-optimal [YRZC20] Bayes optimal undefined outside distribution x Bayes optimal r-optimal r-optimal = classifier that maximizes accuracy at points that have robustness radius at least r

  27. Convergence Result [BC20] Theorem: For r-separated data, condi- tions when non-parametrics converge 2r to r-optimal in large n limit

  28. Convergence Result [BC20] Theorem: For r-separated data, condi- tions when non-parametrics converge 2r to r-optimal in large n limit Convergence limit: r-optimal: Nearest neighbor, Kernel classifiers

  29. Convergence Result [BC20] Theorem: For r-separated data, condi- tions when non-parametrics converge 2r to r-optimal in large n limit Convergence limit: r-optimal: Nearest neighbor, Kernel classifiers Bayes-optimal but not r-optimal: Histograms, Decision trees

  30. Convergence Result [BC20] Theorem: For r-separated data, condi- tions when non-parametrics converge 2r to r-optimal in large n limit Robustness depends on training algorithm! Convergence limit: r-optimal: Nearest neighbor, Kernel classifiers Bayes-optimal but not r-optimal: Histograms, Decision trees

  31. Robustness for r-separated data: Two Settings Non-parametric Methods Neural Networks

  32. Robustness in Neural Networks A large number of attacks A few defenses All defenses show a robustness-accuracy tradeoff Is this tradeoff necessary?

  33. The Setting: Neural Networks f(x) f Neural network computes function f(x) 0 Classifier output sign(f(x)) x

  34. The Setting: Neural Networks f(x) f Neural network computes function f(x) 0 Classifier output sign(f(x)) x Robustness comes from Local Smoothness: If f is locally Lipschitz around x, and f(x) is away from 0, then f is robust at x

  35. Robustness and Accuracy Possible through Local Lipschitzness f(x) f Neural network computes function f(x) 0 Classifier output sign(f(x)) x Theorem [YRZSC20] If distribution is r-separated, then there exists an f s.t. f is locally smooth and sign(f) has accuracy 1 and robustness radius r

  36. In principle, no robustness-accuracy tradeoff In practice there is one What accounts for this gap?

  37. Empirical Study 4 standard image datasets 7 models 6 different training methods - Natural, AT, Trades, LLR, GR Measure local Lipschitzness, accuracy and adversarial accuracy

  38. Result: CIFAR 10

  39. Observations Trades and Adversarial training have best local Lipschitzness Overall, local Lipschitzness correlated with robustness and accuracy - until underfitting begins Generalization gap is quite large - possibly a sign of overfitting Overall: robustness/accuracy tradeoff due to imperfect training methods

  40. Conclusion: Why do we have adversarial examples? - Data Distributional distribution Robustness + Too few - Finite Sample samples Robustness + Algorithmic Bad - Robustness algorithm +

  41. References Robustness for Non-parametric Methods: A generic defense and an attack, Y. Yang, C. Rashtchian, Y. Wang, and K. Chaudhuri, AISTATS 2020 When are Non-parametric Methods Robust? R. Bhattacharjee and K. Chaudhuri, Arxiv 2003.06121 Adversarial Robustness through Local Lipschitzness Y. Yang, C. Rashtchian, H. Zhang, R. Salakhutdinov and K. Chaudhuri, Arxiv 2003.02460

  42. Acknowledgements Cyrus Yaoyuan Yizhen Hongyang Robi Ruslan Rashtchian Yang Wang Zhang Bhattacharjee Salakhutdinov

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend