Smooth Sensitivity and Sampling Sofya Raskhodnikova Penn State - - PowerPoint PPT Presentation

smooth sensitivity and sampling
SMART_READER_LITE
LIVE PREVIEW

Smooth Sensitivity and Sampling Sofya Raskhodnikova Penn State - - PowerPoint PPT Presentation

Smooth Sensitivity and Sampling Sofya Raskhodnikova Penn State University Joint work with Kobbi Nissim ( Ben Gurion University ) and Adam Smith ( Penn State University ) Our main contributions Starting point: Global sensitivity framework


slide-1
SLIDE 1

Smooth Sensitivity and Sampling

Sofya Raskhodnikova

Penn State University

Joint work with Kobbi Nissim (Ben Gurion University) and Adam Smith

(Penn State University)

slide-2
SLIDE 2

Our main contributions

  • Starting point: Global sensitivity framework [DMNS06]

(Cynthia’s talk)

  • Two new frameworks for private data analysis
  • Greatly expand the types of information that can be

released

2

slide-3
SLIDE 3

Road map

  • I. Introduction
  • Review of global sensitivity framework [DMNS06]
  • Motivation
  • II. Smooth sensitivity framework
  • III. Sample-and-aggregate framework

3

slide-4
SLIDE 4

Model

. . . x1 x2 xn

✯ ✲ q ❥

Trusted agency A

✲ ✛

Compute f(x) A(x) = f(x) + noise

Users Each row is arbitrarily complex data supplied by 1 person. For which functions f can we have:

  • utility: little noise
  • privacy: indistinguishability definition of [DMNS06]

4

slide-5
SLIDE 5

Privacy as indistinguishability [DMNS06]

Two databases are neighbors if they differ in one row. x = . . . x1 x2 xn x′ = . . . x1 x′

2

xn Privacy definition Algorithm A is ε-indistinguishable if

  • for all neighbor databases x, x′
  • for all sets of answers S

Pr[A(x) ∈ S] ≤ (1 + ε) · Pr[A(x′) ∈ S]

5

slide-6
SLIDE 6

Privacy definition: composition If A is ε-indistinguishable on each query, it is εq-indistinguishable on q queries.

. . . x1 x2 xn

✯ ✲ q ❥

ε-indisting. agency A

✲ ✛

Compute f(x) A(x) = f(x) + noise

Users

6

slide-7
SLIDE 7

Global sensitivity framework [DMNS06] Intuition: f can be released accurately when it is insensitive to individual entries x1, . . . , xn.

Global sensitivity GSf = max

neighbors x,x′ f(x) − f(x′).

Example: GSaverage = 1

n if x ∈ [0, 1]n.

Theorem If A(x) = f(x) + Lap

  • GSf

ε

  • then A is ε-indistinguishable.

7

slide-8
SLIDE 8

Instance-Based Noise

Big picture for global sensitivity framework:

  • add enough noise to cover the worst case for f
  • noise distribution depends only on f, not database x

Problem: for some functions that’s too much noise

8

slide-9
SLIDE 9

Instance-Based Noise

Big picture for global sensitivity framework:

  • add enough noise to cover the worst case for f
  • noise distribution depends only on f, not database x

Problem: for some functions that’s too much noise Example: median of x1, . . . , xn ∈ [0, 1] x = 0 · · · 0

n−1 2

0 1 · · · 1

n−1 2

x′ = 0 · · · 0

n−1 2

1 1 · · · 1

n−1 2

median(x) = 0 median(x′) = 1 GSmedian = 1

  • Noise magnitude:

1 ε.

8

slide-10
SLIDE 10

Instance-Based Noise

Big picture for global sensitivity framework:

  • add enough noise to cover the worst case for f
  • noise distribution depends only on f, not database x

Problem: for some functions that’s too much noise Example: median of x1, . . . , xn ∈ [0, 1] x = 0 · · · 0

n−1 2

0 1 · · · 1

n−1 2

x′ = 0 · · · 0

n−1 2

1 1 · · · 1

n−1 2

median(x) = 0 median(x′) = 1 GSmedian = 1

  • Noise magnitude:

1 ε.

Our goal: noise tuned to database x

8

slide-11
SLIDE 11

Road map

  • I. Introduction
  • Review of global sensitivity framework [DMNS06]
  • Motivation
  • II. Smooth sensitivity framework
  • III. Sample-and-aggregate framework

9

slide-12
SLIDE 12

Local sensitivity

Local sensitivity LSf(x) = max

x′: neighbor of x f(x) − f(x′)

Reminder: GSf = max

x

LSf(x) Example: median for 0 ≤ x1 ≤ · · · ≤ xn ≤ 1, odd n

1

r r r r r x1 xn xm−1 xm+1 xm

. . . . . .

median

LSmedian(x) = max(xm − xm−1, xm+1 − xm) Goal: Release f(x) with less noise when LSf(x) is lower.

10

slide-13
SLIDE 13

Local sensitivity

Local sensitivity LSf(x) = max

x′: neighbor of x f(x) − f(x′)

Reminder: GSf = max

x

LSf(x) Example: median for 0 ≤ x1 ≤ · · · ≤ xn ≤ 1, odd n

1

r r r r r x1 xn xm−1 xm+1 xm

. . . . . .

median

new median when x′

1 = 1

LSmedian(x) = max(xm − xm−1, xm+1 − xm) Goal: Release f(x) with less noise when LSf(x) is lower.

10

slide-14
SLIDE 14

Local sensitivity

Local sensitivity LSf(x) = max

x′: neighbor of x f(x) − f(x′)

Reminder: GSf = max

x

LSf(x) Example: median for 0 ≤ x1 ≤ · · · ≤ xn ≤ 1, odd n

1

r r r r r x1 xn xm−1 xm+1 xm

. . . . . .

median

new median when x′

n = 0

new median when x′

1 = 1

LSmedian(x) = max(xm − xm−1, xm+1 − xm) Goal: Release f(x) with less noise when LSf(x) is lower.

10

slide-15
SLIDE 15

Instance-based noise: first attempt

Noise magnitude proportional to LSf(x) instead of GSf? No! Noise magnitude reveals information. Lesson: Noise magnitude must be an insensitive function.

11

slide-16
SLIDE 16

Smooth bounds on local sensitivity

Design sensitivity function S(x)

  • S(x) is an ε-smooth upper bound on LSf(x) if:

– for all x:

S(x) ≥ LSf(x)

– for all neighbors x, x′ :

S(x) ≤ eεS(x′)

✲ ✻

x

LSf(x)

Theorem

If A(x) = f(x) + noise S(x) ε

  • then A is ε′-indistinguishable.

Example: GSf is always a smooth bound on LSf(x)

12

slide-17
SLIDE 17

Smooth bounds on local sensitivity

Design sensitivity function S(x)

  • S(x) is an ε-smooth upper bound on LSf(x) if:

– for all x:

S(x) ≥ LSf(x)

– for all neighbors x, x′ :

S(x) ≤ eεS(x′)

✲ ✻

x

LSf(x) S(x)

Theorem

If A(x) = f(x) + noise S(x) ε

  • then A is ε′-indistinguishable.

Example: GSf is always a smooth bound on LSf(x)

12

slide-18
SLIDE 18

Smooth Sensitivity

Smooth sensitivity S∗

f(x)= max y

  • LSf(y)e−ε·dist(x,y)

Lemma For every ε-smooth bound S: S∗

f(x) ≤ S(x) for all x.

Intuition: little noise when far from sensitive instances

database space high local sensitivity low local sensitivity

13

slide-19
SLIDE 19

Smooth Sensitivity

Smooth sensitivity S∗

f(x)= max y

  • LSf(y)e−ε·dist(x,y)

Lemma For every ε-smooth bound S: S∗

f(x) ≤ S(x) for all x.

Intuition: little noise when far from sensitive instances

database space high local sensitivity low local sensitivity low smooth sensitivity

13

slide-20
SLIDE 20

Computing smooth sensitivity

Example functions with computable smooth sensitivity

  • Median & minimum of numbers in a bounded interval
  • MST cost when weights are bounded
  • Number of triangles in a graph

Approximating smooth sensitivity

  • only smooth upper bounds on LS are meaningful
  • simple generic methods for smooth approximations

– work for median and 1-median in Ld

1

14

slide-21
SLIDE 21

Road map

  • I. Introduction
  • Review of global sensitivity framework [DMNS06]
  • Motivation
  • II. Smooth sensitivity framework
  • III. Sample-and-aggregate framework

15

slide-22
SLIDE 22

New goal

  • Smooth sensitivity framework requires

understanding combinatorial structure of f – hard in general

  • Goal: an automatable transformation from

an arbitrary f into an ε-indistinguishable A – A(x) ≈ f(x) for ”good” instances x

16

slide-23
SLIDE 23

Example: cluster centers

Database entries: points in a metric space.

x

r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜

x′

r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜

  • Comparing sets of centers: Earthmover-like metric
  • Global sensitivity of cluster centers is roughly the

diameter of the space. But intuitively, if clustering is ”good”, cluster centers should be insensitive.

  • No efficient approximation for smooth sensitivity

17

slide-24
SLIDE 24

Example: cluster centers

Database entries: points in a metric space.

x

r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ ✫✪ ✬✩

x′

r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ ✫✪ ✬✩

  • Comparing sets of centers: Earthmover-like metric
  • Global sensitivity of cluster centers is roughly the

diameter of the space. But intuitively, if clustering is ”good”, cluster centers should be insensitive.

  • No efficient approximation for smooth sensitivity

17

slide-25
SLIDE 25

Example: cluster centers

Database entries: points in a metric space.

x

r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ ✉ ❡ ✉ ❡ ✫✪ ✬✩

x′

r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ r ❜ ✉ ❡ ✉ ❡ ✫✪ ✬✩

  • Comparing sets of centers: Earthmover-like metric
  • Global sensitivity of cluster centers is roughly the

diameter of the space. But intuitively, if clustering is ”good”, cluster centers should be insensitive.

  • No efficient approximation for smooth sensitivity

17

slide-26
SLIDE 26

Sample-and-aggregate framework

Intuition: Replace f with a less sensitive function ˜ f. ˜ f(x) = g(f(sample1), f(sample2), . . . , f(samples))

✮ ☛ q ❄ ❄ ❄ ❥◆ ✙

x

xi1, . . . , xit xj1, . . . , xjt

. . .

xk1, . . . , xkt ⑥ ⑥ ⑥ ⑥ ♠ ♠ ♠ ♠ f f f

g

aggregation function

18

slide-27
SLIDE 27

Sample-and-aggregate framework

Intuition: Replace f with a less sensitive function ˜ f. ˜ f(x) = g(f(sample1), f(sample2), . . . , f(samples))

✮ ☛ q ❄ ❄ ❄ ❥◆ ✙

x

xi1, . . . , xit xj1, . . . , xjt

. . .

xk1, . . . , xkt ⑥ ⑥ ⑥ ⑥ ♠ ♠ ♠ ♠ f f f

g

aggregation function ❄ ✲ ✲ ♠

+

noise calibrated to sensitivity of ˜ f

  • utput

18

slide-28
SLIDE 28

Good aggregation functions

  • average

– works for L1 and L2

  • center of attention

– the center of a smallest ball containing a strict majority of input points – works for arbitrary metrics (in particular, for Earthmover) – gives lower noise for L1 and L2

19

slide-29
SLIDE 29

Sample-and-aggregate results

Theorem If f can be approximated on x from small samples then f can be released with little noise

20

slide-30
SLIDE 30

Sample-and-aggregate results

Theorem If f can be approximated on x within distance r from small samples of size n1−δ then f can be released with little noise ≈ r

ε + negl(n)

20

slide-31
SLIDE 31

Sample-and-aggregate results

Theorem If f can be approximated on x within distance r from small samples of size n1−δ then f can be released with little noise ≈ r

ε + negl(n)

  • Works in all ”interesting” metric spaces
  • Example applications

– k-means cluster centers (if data is separated a.k.a. [Ostrovsky Rabani Schulman Swamy 06]) – fitting mixtures of Gaussians (if data is i.i.d., using [Vempala Wang 04, Achlioptas McSherry 05]) – PAC concepts (Adam Smith’s talk)

20

slide-32
SLIDE 32

Road map

  • I. Introduction
  • Review of global sensitivity framework [DMNS06]
  • Motivation
  • II. Smooth sensitivity framework
  • III. Sample-and-aggregate framework

21

slide-33
SLIDE 33

Conclusion: fundamental question

Which computations are not too sensitive to individual inputs?

Which functions f admit ε-indistinguishable approximation A?

22