Adversarial Regression with Multiple Learners Liang Tong 1 Sixie Yu - - PowerPoint PPT Presentation

adversarial regression with multiple learners
SMART_READER_LITE
LIVE PREVIEW

Adversarial Regression with Multiple Learners Liang Tong 1 Sixie Yu - - PowerPoint PPT Presentation

Adversarial Regression with Multiple Learners Liang Tong 1 Sixie Yu 1 Scott Alfeld 2 Yevgeniy Vorobeychik 1 1 Electrical Engineering and Computer Science Vanderbilt University 2 Computer Science Amherst College ICML 2018 ( Electrical


slide-1
SLIDE 1

Adversarial Regression with Multiple Learners

Liang Tong∗1 Sixie Yu∗1 Scott Alfeld2 Yevgeniy Vorobeychik1

1Electrical Engineering and Computer Science

Vanderbilt University

2Computer Science

Amherst College

ICML 2018

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 1 / 21

slide-2
SLIDE 2

Problem Setting

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 2 / 21

slide-3
SLIDE 3

Motivation

Adversaries can change features at test time to cause incorrect predictions.

i.e., change features of a house (i.e., square feet, #rooms) to fool

  • nline real-estate evaluation system, or make invisible changes to

pictures to fool classifier.

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 3 / 21

slide-4
SLIDE 4

Motivation

Adversaries can change features at test time to cause incorrect predictions.

i.e., change features of a house (i.e., square feet, #rooms) to fool

  • nline real-estate evaluation system, or make invisible changes to

pictures to fool classifier.

Previous investigations of this problem pit a single learner against an

  • adversary. [Bruckner11, Dalvi04, li2014feature, zhou2012 ]

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 3 / 21

slide-5
SLIDE 5

Motivation

Adversaries can change features at test time to cause incorrect predictions.

i.e., change features of a house (i.e., square feet, #rooms) to fool

  • nline real-estate evaluation system, or make invisible changes to

pictures to fool classifier.

Previous investigations of this problem pit a single learner against an

  • adversary. [Bruckner11, Dalvi04, li2014feature, zhou2012 ]

But an adversary’s decision is usually aimed at a collection of learners.

i.e., an adversary crafts generic malwares and disseminate them widely.

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 3 / 21

slide-6
SLIDE 6

Motivation

Adversaries can change features at test time to cause incorrect predictions.

i.e., change features of a house (i.e., square feet, #rooms) to fool

  • nline real-estate evaluation system, or make invisible changes to

pictures to fool classifier.

Previous investigations of this problem pit a single learner against an

  • adversary. [Bruckner11, Dalvi04, li2014feature, zhou2012 ]

But an adversary’s decision is usually aimed at a collection of learners.

i.e., an adversary crafts generic malwares and disseminate them widely.

The learners all make autonomous decisions about how to detect malicious content.

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 3 / 21

slide-7
SLIDE 7

Table of Contents

1

Learner Model

2

Attacker Model

3

Multi-Learner Stackelberg Game (MLSG)

4

Existence and Uniqueness of the Equilibrium

5

Computing the MLNE

6

Robustness analysis

7

References

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 4 / 21

slide-8
SLIDE 8

Learner Model

(X, y): training dataset from an unknown distribution D. X = [x1, ..., xm]⊤ and y = [y1, y2, ..., ym]⊤: xj the jth instance and yj its corresponding response variable. Test data is drawn from a distribution D

′ (a modification of D)

manipulated by the attacker. An instance from D

′ (D) with probability β (1 − β).

The action of the ith learner is to learn the parameters of the linear regression model: θi, which results in ˆ yi = Xθi. The expected cost function of the ith learner: ci(θi, D

′) = βE(X′,y)∼D′[ℓ(X ′θi, y)] + (1 − β)E(X,y)∼D[ℓ(Xθi, y)]

(1) where ℓ(ˆ y, y) = ||ˆ y − y||2

2.

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 5 / 21

slide-9
SLIDE 9

Table of Contents

1

Learner Model

2

Attacker Model

3

Multi-Learner Stackelberg Game (MLSG)

4

Existence and Uniqueness of the Equilibrium

5

Computing the MLNE

6

Robustness analysis

7

References

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 6 / 21

slide-10
SLIDE 10

Attacker Model

Every instance (x, y) is maliciously modified by the attacker to (x

′, y),

with probability β. Assume the attacker has an instance-specific target z(x). The objective of the attacker: ˆ y = θ⊤

i x

′ close to z(x).

The attacker’s objective is measured by: ℓ(ˆ y, z) = ||ˆ y − z||2

2.

Transforming X to X

′ incurs costs: R(X ′, X) = ||X ′ − X||2

F.

The expected cost function of the attacker: ca({θi}n

i=1, X

′) =

n

  • i=1

ℓ(X

′θi, z) + λR(X ′, X)

(2)

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 7 / 21

slide-11
SLIDE 11

Table of Contents

1

Learner Model

2

Attacker Model

3

Multi-Learner Stackelberg Game (MLSG)

4

Existence and Uniqueness of the Equilibrium

5

Computing the MLNE

6

Robustness analysis

7

References

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 8 / 21

slide-12
SLIDE 12

Multi-Learner Stackelberg Game (MLSG)

The MLSG has two stages, which proceeds as follow:

In the first stage the learners simultaneously learn their model parameters {θi}n

i=1.

In the second stage, after observing the learners’ decision, the attacker constructs its optimal attack (manipulating X).

Assumptions

The learners have complete information about β, λ, and z. Each learner has the same action space Θ ⊆ Rd×1, which is nonempty, compact, and convex. The columns of the test data X are linearly independent.

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 9 / 21

slide-13
SLIDE 13

Multi-Learner Stackelberg Game (MLSG)

Definition (Multi-Learner Stackelberg Equilibrium (MLSE))

An action profile ({θ∗

i }n i=1, X∗) is an MLSE if it satisfies

θ∗

i = arg min θi∈Θ

ci(θi, X∗(θ)), ∀i ∈ N s.t. X∗(θ) = arg min

X′∈Rm×d

ca({θi}n

i=1, X

′).

(3) where θ = {θi}n

i=1 constitutes the joint actions of the learners.

MLSE is a blend between a Nash equilibrium (among all learners) and a Stackelberg equilibrium (between the learners and the attacker).

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 10 / 21

slide-14
SLIDE 14

Multi-Learner Stackelberg Game (MLSG)

Lemma (Best Response of the Attacker)

Given {θi}n

i=1, the best response of the attacker is

X∗ = (λX + z

n

  • i=1

θ⊤

i )(λI + n

  • i=1

θiθ⊤

i )−1.

(4) X∗ has a closed form, as a function of {θi}n

i=1.

With this lemma, the learners’ cost functions become: ci(θi, θ−i) = βℓ(X∗(θi, θ−i)θi, y) + (1 − β)ℓ(Xθi, y). (5) MLSG

X∗(θi,θ−i)

= = = = = = ⇒ Multi-Learner Nash Game (MLNG) MLNG is a game among the learners.

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 11 / 21

slide-15
SLIDE 15

Table of Contents

1

Learner Model

2

Attacker Model

3

Multi-Learner Stackelberg Game (MLSG)

4

Existence and Uniqueness of the Equilibrium

5

Computing the MLNE

6

Robustness analysis

7

References

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 12 / 21

slide-16
SLIDE 16

Existence and Uniqueness of the Equilibrium

We approximate the MLNG by deriving upper bounds on the learners’ cost

  • functions. The approximated game is denoted by: N, Θ, (

ci).

Theorem (Existence of Nash Equilibrium)

N, Θ, ( ci) is a symmetric game and it has at least one symmetric equilibrium.

Theorem (Uniqueness of Nash Equilibrium)

N, Θ, ( ci) has an unique Nash equilibrium, and this unique NE is symmetric. The equilibrium of N, Θ, ( ci) is defined as: Multi-Learner Nash Equilibrium (MLNE)

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 13 / 21

slide-17
SLIDE 17

Table of Contents

1

Learner Model

2

Attacker Model

3

Multi-Learner Stackelberg Game (MLSG)

4

Existence and Uniqueness of the Equilibrium

5

Computing the MLNE

6

Robustness analysis

7

References

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 14 / 21

slide-18
SLIDE 18

Computing the MLNE

By utilizing first-order optimality conditions of each learner’s optimization problem:

Theorem

Let f (θ) = ℓ(Xθ, y) + β(n + 1) 2λ2 ||z − y||2

2(θ⊤θ)2,

(6) Then, the unique symmetric NE of N, Θ, ( ci), {θ∗

i }n i=1, can be derived

by solving the following convex optimization problem min

θ∈Θ f (θ)

(7) and then letting θ∗

i = θ∗, ∀i ∈ N, where θ∗ is the solution of Eq. (7).

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 15 / 21

slide-19
SLIDE 19

Table of Contents

1

Learner Model

2

Attacker Model

3

Multi-Learner Stackelberg Game (MLSG)

4

Existence and Uniqueness of the Equilibrium

5

Computing the MLNE

6

Robustness analysis

7

References

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 16 / 21

slide-20
SLIDE 20

Robustness analysis

A robust linear regression solves the following problem: min

θ∈Θ max △∈U ||y − (X + △)θ||2 2,

(8) where the uncertainty set U = {△ ∈ Rm×d | △T△ = G : |Gij| ≤ c|θiθj| ∀i, j}, with c = β(n+1)

2λ2

||z − y||2

2.

Theorem

The optimal solution θ∗ of the problem in Eq. (7) is an optimal solution to the robust optimization problem in Eq. (8).

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 17 / 21

slide-21
SLIDE 21

Our Contribution

Fomally model the interaction between the learners and the attacker as a Multi-Learner Stackelberg Game.

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 18 / 21

slide-22
SLIDE 22

Our Contribution

Fomally model the interaction between the learners and the attacker as a Multi-Learner Stackelberg Game. Approximate this game by deriving upper bounds on the learners’ loss functions.

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 18 / 21

slide-23
SLIDE 23

Our Contribution

Fomally model the interaction between the learners and the attacker as a Multi-Learner Stackelberg Game. Approximate this game by deriving upper bounds on the learners’ loss functions. Show that there always exists a unique symmetric equilibrium of the approximated game.

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 18 / 21

slide-24
SLIDE 24

Our Contribution

Fomally model the interaction between the learners and the attacker as a Multi-Learner Stackelberg Game. Approximate this game by deriving upper bounds on the learners’ loss functions. Show that there always exists a unique symmetric equilibrium of the approximated game. Theoretically and experimently show that the equilibrium of the approximated game is robust.

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 18 / 21

slide-25
SLIDE 25

Thank you! Poster: Hall B #120 Email: sixie.yu@vanderbilt.edu Homepage: sixie-yu.org

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 19 / 21

slide-26
SLIDE 26

Table of Contents

1

Learner Model

2

Attacker Model

3

Multi-Learner Stackelberg Game (MLSG)

4

Existence and Uniqueness of the Equilibrium

5

Computing the MLNE

6

Robustness analysis

7

References

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 20 / 21

slide-27
SLIDE 27

References

( Electrical Engineering and Computer Science Vanderbilt University, Computer Science Amherst College ) Adversarial Regression with Multiple Learners ICML 2018 21 / 21