Exploring the Landscape of Spa5al Robustness Logan Engstrom (with - - PowerPoint PPT Presentation

exploring the landscape of spa5al robustness
SMART_READER_LITE
LIVE PREVIEW

Exploring the Landscape of Spa5al Robustness Logan Engstrom (with - - PowerPoint PPT Presentation

Exploring the Landscape of Spa5al Robustness Logan Engstrom (with Brandon Tran*, Dimitris Tsipras*, Ludwig Schmidt, Aleksander Mdry) madry-lab.ml ML Glitch: Adversarial Examples ML Glitch: Adversarial Examples pig small,


slide-1
SLIDE 1

Exploring the Landscape

  • f Spa5al Robustness

Logan Engstrom

(with Brandon Tran*, Dimitris Tsipras*, Ludwig Schmidt, Aleksander Mądry) madry-lab.ml

slide-2
SLIDE 2

ML “Glitch”: Adversarial Examples

slide-3
SLIDE 3

“pig” “airliner” small, nonrandom noise

ML “Glitch”: Adversarial Examples

slide-4
SLIDE 4

“pig” “airliner” small, non-random noise

ML “Glitch”: Adversarial Examples

slide-5
SLIDE 5

“pig” “airliner” small, non-random noise

ML “Glitch”: Adversarial Examples

slide-6
SLIDE 6

“pig” “airliner” small, non-random noise

What does small mean here?

ML “Glitch”: Adversarial Examples

slide-7
SLIDE 7

“pig” “airliner” small, non-random noise

What does small mean here?

Traditionally: perturbations that have small l_p norm

ML “Glitch”: Adversarial Examples

slide-8
SLIDE 8

“pig” “airliner” small, non-random noise

What does small mean here?

Traditionally: perturbations that have small l_p norm Do small l_p norms capture every sense of “small”?

ML “Glitch”: Adversarial Examples

slide-9
SLIDE 9

Spa5al Perturba5ons

slide-10
SLIDE 10

Spa5al Perturba5ons

slide-11
SLIDE 11

Spa5al Perturba5ons

rotation up to 30°

slide-12
SLIDE 12

Spa5al Perturba5ons

rotation up to 30° x, y translations up to ~10%

slide-13
SLIDE 13

Spa5al Perturba5ons

rotation up to 30° x, y translations up to ~10%

These are not small l_p perturbations!

slide-14
SLIDE 14

Spa5al Perturba5ons

rotation up to 30° x, y translations up to ~10%

How robust are models to spatial perturbations? These are not small l_p perturbations!

slide-15
SLIDE 15

Spa5al Robustness

slide-16
SLIDE 16

Spa5al Robustness

Spoiler: models are not robust

slide-17
SLIDE 17

Spa5al Robustness

Spoiler: models are not robust

slide-18
SLIDE 18

Spa5al Robustness

Spoiler: models are not robust Can we train more spatially robust classifiers?

slide-19
SLIDE 19

Spa5al Defenses

slide-20
SLIDE 20

Spa5al Defenses

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

slide-21
SLIDE 21

Spa5al Defenses

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

Key question: how to find worst-case translations, rotations?

slide-22
SLIDE 22

Attempt #1: first-order methods

Spa5al Defenses

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

Key question: how to find worst-case translations, rotations?

slide-23
SLIDE 23

Attempt #1: first-order methods

Spa5al Defenses

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

Key question: how to find worst-case translations, rotations?

slide-24
SLIDE 24

Attempt #1: first-order methods

Spa5al Defenses

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

Key question: how to find worst-case translations, rotations?

slide-25
SLIDE 25

Spa5al Defenses

Key question: how to find worst-case translations, rotations?

Attempt #1: first-order methods

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

slide-26
SLIDE 26

Spa5al Defenses

Key question: how to find worst-case translations, rotations?

Attempt #1: first-order methods Attempt #2: exhaustive search

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

slide-27
SLIDE 27

Spa5al Defenses

Key question: how to find worst-case translations, rotations?

Attempt #1: first-order methods Attempt #2: exhaustive search

Exhaustive search is feasible, and a strong adversary! (discretize translations and rotations, try every combination)

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

slide-28
SLIDE 28

Spa5al Defenses

Key question: how to find worst-case translations, rotations?

Attempt #1: first-order methods Attempt #2: exhaustive search

Exhaustive search is feasible, and a strong adversary! (discretize translations and rotations, try every combination)

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

slide-29
SLIDE 29

Spa5al Defenses

Key question: how to find worst-case translations, rotations?

Attempt #1: first-order methods Attempt #2: exhaustive search

Train only on “worst” transformed input (highest loss)

Exhaustive search is feasible, and a strong adversary! (discretize translations and rotations, try every combination)

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

slide-30
SLIDE 30

Spa5al Defenses

Key question: how to find worst-case translations, rotations?

Attempt #1: first-order methods Attempt #2: exhaustive search

Exhaustive search is feasible, and a strong adversary! (discretize translations and rotations, try every combination) (we approximate via 10 random samples to quicken training)

Lesson from l_p robustness: use robust optimization

[Goodfellow et al ‘15 ][Madry et al ’18]

(= train on worst-case perturbed inputs)

slide-31
SLIDE 31

Spa5al Defenses

With robust optimization:

slide-32
SLIDE 32

Spa5al Defenses

CIFAR classifier accuracy: 3% adversarial to 71% adversarial With robust optimization:

slide-33
SLIDE 33

Spa5al Defenses

CIFAR classifier accuracy: 3% adversarial to 71% adversarial (compare to 93% standard accuracy) With robust optimization:

slide-34
SLIDE 34

Spa5al Defenses

ImageNet classifier accuracy: 31% adversarial to 53% adversarial CIFAR classifier accuracy: 3% adversarial to 71% adversarial (compare to 93% standard accuracy) With robust optimization:

slide-35
SLIDE 35

Spa5al Defenses

ImageNet classifier accuracy: 31% adversarial to 53% adversarial (compare to 76% standard accuracy) CIFAR classifier accuracy: 3% adversarial to 71% adversarial (compare to 93% standard accuracy) With robust optimization:

slide-36
SLIDE 36

Spa5al Defenses

With robust optimization: ImageNet classifier accuracy: 31% adversarial to 53% adversarial (compare to 76% standard accuracy) CIFAR classifier accuracy: 3% adversarial to 71% adversarial (compare to 93% standard accuracy) (+10 sample majority vote)

slide-37
SLIDE 37

Spa5al Defenses

With robust optimization: ImageNet classifier accuracy: 31% adversarial to 53% adversarial (compare to 76% standard accuracy) CIFAR classifier accuracy: 3% adversarial to 71% adversarial (compare to 93% standard accuracy) (+10 sample majority vote) 82%

slide-38
SLIDE 38

Spa5al Defenses

With robust optimization: ImageNet classifier accuracy: 31% adversarial to 53% adversarial (compare to 76% standard accuracy) CIFAR classifier accuracy: 3% adversarial to 71% adversarial (compare to 93% standard accuracy) (+10 sample majority vote) 82% 56%

slide-39
SLIDE 39

Spa5al Defenses

With robust optimization: Still significant room for improvement! ImageNet classifier accuracy: 31% adversarial to 53% adversarial (compare to 76% standard accuracy) CIFAR classifier accuracy: 3% adversarial to 71% adversarial (compare to 93% standard accuracy) (+10 sample majority vote) 82% 56%

slide-40
SLIDE 40

Conclusions

slide-41
SLIDE 41

Conclusions

Robust models need more refined notions of similarity

slide-42
SLIDE 42

Conclusions

We do not have true spatial robustness Robust models need more refined notions of similarity

slide-43
SLIDE 43

Conclusions

Intuitions from l_p robustness do not transfer We do not have true spatial robustness Robust models need more refined notions of similarity

slide-44
SLIDE 44

Conclusions

Come to our poster! Pacific Ballroom #142 Intuitions from l_p robustness do not transfer We do not have true spatial robustness Robust models need more refined notions of similarity