Adversarial Learning Bounds for Linear Classes and Neural Nets - - PowerPoint PPT Presentation

adversarial learning bounds for linear classes and neural
SMART_READER_LITE
LIVE PREVIEW

Adversarial Learning Bounds for Linear Classes and Neural Nets - - PowerPoint PPT Presentation

Adversarial Learning Bounds for Linear Classes and Neural Nets Understanding Adversarial Learning through Rademacher Complexity Pranjal Awasthi, Natalie Frank, Mehryar Mohri Google Research & Courant Institute August 14, 2020 1 / 17


slide-1
SLIDE 1

Adversarial Learning Bounds for Linear Classes and Neural Nets

Understanding Adversarial Learning through Rademacher Complexity Pranjal Awasthi, Natalie Frank, Mehryar Mohri

Google Research & Courant Institute

August 14, 2020

1 / 17

slide-2
SLIDE 2

Adversarial Attacks

Figure: Imperceptible adversarial perturbations in ℓ2. [5]

2 / 17

slide-3
SLIDE 3

Adversarial Robustness

Figure: A sparse perturbation. [1]

Overarching Goal: Train classifiers robust to adversarial perturbations. ◮ Examples in many areas of applications. ◮ Different possible forms of perturbations: changing every pixel in an image vs. placing a sticker on a stop sign. ◮ Can we derive learning guarantees for adversarial robustness?

3 / 17

slide-4
SLIDE 4

Outline of Talk

Goal of our paper: Understand what characterizes robust generalization and how it relates to non-robust generalization

  • 1. Classification & Adversarial Classification setup
  • 2. Rademacher complexity & Adversarial Rademacher

Complexity

  • 3. Better bounds for adversarial Rademacher complexity of linear

classes

  • 4. Better bounds for Rademacher complexity of linear classes
  • 5. Adversarial Rademacher complexity of neural nets

4 / 17

slide-5
SLIDE 5

Standard Classification Setting

Binary Classification: Data distributed over Rd × {−1, +1} according to D Standard Setting: ◮ Given a predictor h : Rd → R, a point x is classified as sign(h(x)). ◮ There is an error if yh(x) < 0, or 1yh(x)<0 = 1. ◮ The classification error is then R(h) = E

(x,y)∼D[1yh(x)<0]

5 / 17

slide-6
SLIDE 6

Defining Adversarial Perturbations

Adversarial Setting: ◮ The data is perturbed by ǫ in ℓp to “fool” the classifier into thinking there is an error, now an error occurs if 1 = sup

x−x′r≤ǫ

1yh(x′)<0 = 1infx−x′r ≤ǫ yh(x′)<0 ◮ The adversarial classification error is then

  • R(h) =

E

(x,y)∼D[1infx−x′r ≤ǫ yh(x′)<0]

6 / 17

slide-7
SLIDE 7

Rademacher Complexity

The empirical Rademacher complexity is RS(F) = E

σ

  • sup

f ∈F

1 m

m

  • i=1

σif (zi)

  • Theorem (Margin Bounds [4])

R(h) ≤ RS,ρ(h)+2 ρRS(F)+3

  • log 2

δ

2m holds with probability at least 1 − δ for all h ∈ F. ρ-Margin Loss: Φρ(u) = min(1, max(0, 1 − u ρ))

1 1 7 / 17

slide-8
SLIDE 8

Adversarial Rademacher Complexity

Theorem (Robust margin bounds)

Define the class F by

  • F =
  • (x, y) →

inf

x−x′r≤ǫ yf (x′): f ∈ F

  • .

The following holds with probability at least 1 − δ for all h ∈ F:

  • R(h) ≤

RS,ρ(h) + 2 ρRS( F) + 3

  • log 2

δ

2m .

Definition

We define the adversarial Rademacher Complexity as

  • RS(F) := RS(

F)

8 / 17

slide-9
SLIDE 9

Prior Work on Adversarial Rademacher Complexity of Linear Classes

Fp = {x → w, x: wp ≤ W } Yin et. al. [6]: For perturbations in the infinity norm, for some constant c max(RS(Fp), cǫW d

1 p∗

√m) ≤ RS(Fp) ≤ RS(Fp) + ǫW d

1 p∗

√m Khim and Loh [3]: For perturbation in the r-norm, there exists a constant Mr for which RS(F2) ≤ W √m max

(xi,yi)∈S xi2 + ǫ Mr∗

2√m

9 / 17

slide-10
SLIDE 10

Adversarial Rademacher Complexity of Linear Classes

Fp = {x → w, x: wp ≤ W }

Theorem

Let ǫ > 0, r ≥ 1. Consider a sample S = {(x1, y1), . . . , (xm, ym)} with xi ∈ Rd and yi ∈ {±1} and perturbations in the r-norm. Then max

  • RS(Fp), ǫW max(d1− 1

r − 1 p , 1)

2 √ 2m

RS(Fp) ≤ RS(Fp) + ǫ W 2√m max(d1− 1

r − 1 p , 1) 10 / 17

slide-11
SLIDE 11

Rademacher Complexity of Linear Classes

Fp = {x → w, x: wp ≤ W } X = [x1 . . . xm] Group norms: Ap,q = (A1p · · · Amp)q where Ai is the ith row of A. Prior Work [2]: RS(Fp) ≤

  • W
  • 2 log(2d)

m

Xmax if p = 1

W m

√p∗ − 1Xp∗,2 if 1 < p ≤ 2 Our new bounds: RS(Fp) ≤         

W m

  • 2 log(2d)XT2,p∗

if p = 1

√ 2W m

  • Γ( p∗+1

2 ) √π

1

p∗

XT2,p∗ if 1 < p ≤ 2

W m XT2,p∗

if p ≥ 2

11 / 17

slide-12
SLIDE 12

Comparing the Bounds for 1 < p ≤ 2

RS(Fp) ≤     

W m

√p∗ − 1Xp∗,2

  • ld bound

√ 2W m

  • Γ( p∗+1

2 ) √π

1

p∗

XT2,p∗ new bound Comparing the Norms: If p ≤ 2, then min(m, d)

1 2 − 1 p∗ XT2,p∗ ≥ Xp∗,2 ≥ XT2,p∗

Comparing the Constants: c1(p) =

  • p∗ − 1

c2(p) = √ 2 Γ( p∗+1

2

) √π 1

p∗

5 10 15 20 25 1 1.5 2 2.5 3 3.5 4 4.5 5 5.5

12 / 17

slide-13
SLIDE 13

Adversarial Rademacher Complexity of the ReLU

Gp = {(x, y) → (yw, x)+ : wp ≤ W , y ∈ {−1, 1}} Fp = {x → w, x: wp ≤ W }

Theorem

The adversarial Rademacher complexity of Gp can be bounded as follows: W δǫ 2 √ 2m |T δ

ǫ,s∗| max(d1− 1

p − 1 r , 1) ≤

RS(Gp) ≤ RTǫ(Fp) + ǫ W 2√m max(1, d1− 1

r − 1 p ),

where Tǫ = {i : yi = −1 or , yi = 1 and xir > ǫ} T δ

ǫ,s = {i : s, xi − (1 + δyi)yiǫsr∗>0}

and s∗ is the adversarial perturbation.

13 / 17

slide-14
SLIDE 14

Adversarial Rademacher Complexity of Neural Nets

Gn

p =

  • (x, y) → y

n

  • j=1

ujρ(wj · x): u1 ≤ Λ, wjp ≤ W

  • .

Theorem

Let ρ be a function with Lipschitz constant Lρ with ρ(0) = 0. Then, the following upper bound holds for the adversarial Rademacher complexity of Gn

p:

  • RS(Gn

p) ≤ Lρ

W Λ max(1, d1− 1

p − 1 r )(Xr,∞ + ǫ)

√m

  • ×
  • 1 +
  • d(n + 1) log(36)
  • .

14 / 17

slide-15
SLIDE 15

Towards Dimension Independent Bounds

◮ Studying the structure of adversarial perturbations leads to equations qualitatively similar to γ-fat shattering. ◮ Under appropriate assumptions, this can lead to dimension independent bounds.

15 / 17

slide-16
SLIDE 16

Conclusion

We covered ◮ New bounds for Rademacher complexity of linear classes. ◮ New bounds for adversarial Rademacher complexity of linear classes. ◮ New bounds for adversarial Rademacher complexity of Neural nets. Open problems ◮ Generalize to arbitrary norms: in general is the dual norm a good regularizer? ◮ Improve the adversarial neural nets generalization bound or find a matching lower bound.

16 / 17

slide-17
SLIDE 17

Bibliography

[1] Tianyu Gu, Brendan Dolan-Gavitt, and Siddharth Garg. Badnets: Identifying vulnerabilities in the machine learning model supply chain. CoRR, 2017. [2] Sham M. Kakade, Karthik Sridharan, and Ambuj Tewari. On the complexity of linear prediction: Risk bounds, margin bounds, and regularization. In Proceedings of NIPS, pages 793–800, 2008. [3] Justin Khim and Po-Ling Loh. Adversarial risk bounds via function transformation. arXiv preprint arXiv:1810.09519, 2018. [4] Mehryar Mohri, Afshin Rostamizadeh, and Ameet Talwalkar. Foundations of Machine Learning. The MIT Press, second edition, 2018. [5] Christian Szegedy, Wojciech Zaremba, Ilya Sutskever, Joan Bruna, Dumitru Erhan, Ian J. Goodfellow, and Rob Fergus. Intriguing properties of neural networks. In Proceedings of ICLR, 2014. [6] Dong Yin, Kannan Ramchandran, and Peter L. Bartlett. Rademacher complexity for adversarially robust generalization. In Proceedings of ICML, pages 7085–7094, 2019.

17 / 17