L EARNING UNDER D IFFERENTIAL P RIVACY K OBBI N ISSIM BGU/Harvard - - PowerPoint PPT Presentation

l earning under
SMART_READER_LITE
LIVE PREVIEW

L EARNING UNDER D IFFERENTIAL P RIVACY K OBBI N ISSIM BGU/Harvard - - PowerPoint PPT Presentation

L EARNING UNDER D IFFERENTIAL P RIVACY K OBBI N ISSIM BGU/Harvard Caltech, Spring 2015 Based on joint work with: Amos Beimel, Avrim Blum, Hai Brenner, Mark Bun, Cynthia Dwork, Shiva Kasiviswanathan, Homin Lee, Frank McSherry, Sofya


slide-1
SLIDE 1

KOBBI NISSIM

BGU/Harvard

Caltech, Spring 2015 Based on joint work with: Amos Beimel, Avrim Blum, Hai Brenner, Mark Bun, Cynthia Dwork, Shiva Kasiviswanathan, Homin Lee, Frank McSherry, Sofya Raskhodnikova, Adam Smith, Uri Stemmer, and Salil Vadhan.

LEARNING UNDER DIFFERENTIAL PRIVACY

slide-2
SLIDE 2

LET’S DO SOME SWEET SCIENCE!

— Scurvy: a problem throughout human history — Caused by vitamin C deficiency — How much vitamin C is enough?

Thanks: Mark Bun

slide-3
SLIDE 3

SO YOU HAVE SOME DATA…

2 24 57 83 121 153 176 182 . . . . . .

0 1 2 3 . . . . . . T

Vitamin C level Thanks: Mark Bun

slide-4
SLIDE 4

SO YOU HAVE SOME DATA…

. . . . . .

0 1 2 3 . . . . . . T

2 24 57 83 121 153 176 182 Vitamin C level Thanks: Mark Bun x x x x

— c: threshold func that is consistent with data

slide-5
SLIDE 5

SO YOU HAVE SOME DATA…

Thanks: Mark Bun . . . . . .

0 1 2 3 . . . . . . T

2 24 57 83 121 153 176 182 Vitamin C level x x x x

— c: threshold func that is consistent with data

slide-6
SLIDE 6

SO YOU HAVE SOME DATA…

— c: threshold func that is consistent with data — Theorem: if n > n0 then c also “agrees” with

underlying distribution

¡ n0 depends on learner accuracy and success probability ¡ n0 examples suffice independent of domain size!

Thanks: Mark Bun . . . 2 24 57 83 121 153 x x x x

slide-7
SLIDE 7

WHAT’S THE PROBLEM?

— The hypothesis threshold reveals someone’s data

point!

— With the right auxiliary information, could be

linked to Shiva!

Thanks: Mark Bun . . . 2 24 57 83 121 153 x x x x

slide-8
SLIDE 8

SAVING SHIVA’S PRIVACY

— Idea: “noisy” choice of threshold hides individual

contribution!

Thanks: Mark Bun . . . . . .

0 1 2 3 . . . . . . T

2 24 57 83 121 153 176 182 Vitamin C level x x x x

slide-9
SLIDE 9

Had it been the year 2000 (AD) …

slide-10
SLIDE 10

Had it been the year 2000 (AD) …

slide-11
SLIDE 11

THANK YOU!

Brilliant, isn’t it? Time for coffee and cookies

? ? ?

But ...

slide-12
SLIDE 12

LET’S TAKE A STEP BACK

slide-13
SLIDE 13

DATA PRIVACY – THE PROBLEM

— Given:

¡ A dataset with sensitive information

— How to:

¡ Compute and release functions of the dataset without

compromising individual privacy

A

queries answers )

(

Government, researchers, businesses (or) Malicious adversary

Server/agency Users Database X

xn xn-1 x3 x2 x1

slide-14
SLIDE 14

DATA PRIVACY – THE PROBLEM

— Given:

¡ A dataset with sensitive information

— How to:

¡ Compute and release functions of the dataset without

compromising individual privacy

— Hospital: (based on past patients) predict whether a patient

is prone to scurvy, based on vitamin c level in her blood

— Bank: (based on past customers) predict whether a new

customer is good/bad credit, based on her attributes

— Example, label, presence in database may all be sensitive!

slide-15
SLIDE 15

Differential Privacy [DMNS 06]

— Evolved in [DN’03, EGS’03, DN’04, BDMN’05, DMNS’06, DKMMN’06] — Intuition: to protect an individual make sure

that changing her record does not change the

  • utput distribution (by too much)

xn xn-1 x’3 x2 x1 A A(D’) D’

  • xn

xn-1 x3 x2 x1 A A(D)

  • D
slide-16
SLIDE 16

Differential Privacy [DMNS 06]

Definition: A algorithm A is (ε,δ)-differentially private if: for all neighboring databases D, D’ and for all sets of answers S:

Pr[A(D) ∈ S] ≤ eε Pr[A(D’) ∈ S] + δ

xn xn-1 x3 x2 x1 A A(D)

  • D

xn xn-1 x’3 x2 x1 A A(D’) D’

slide-17
SLIDE 17

Differential Privacy [DMNS 06]

Definition: A algorithm A is (ε,δ)-differentially private if: for all neighboring databases D, D’ and for all sets of answers S:

Pr[A(D) ∈ S] ≤ eε Pr[A(D’) ∈ S] + δ

𝜀≪​1/𝑜

Pure: 𝜀=0 Approx.: 𝜀>0

​𝑓↑𝜗 ≈1+𝜗. . Take 𝜗>​1/𝑜 ¡,

  • therwise, no

utility!

slide-18
SLIDE 18

LEARNING

slide-19
SLIDE 19

PAC Model [Valiant 84]

Samples drawn according to P

Fresh point picked according to distribution P With high probability (over randomness of learner and distribution), a random point drawn according to P is “classified” correctly

A distribution P on X. Each point in X labeled 0/1

slide-20
SLIDE 20

PAC Learning: Definition

— Given distribution P over examples, labeled by c — Hypothesis h is α-­‑good if error(h) = Prx~P [h(x) ≠ c(x)] ≤ α — C: a set of concepts {c: {0,1}d→{0, 1}} — H: a set of hypotheses {h: {0,1}d→{0, 1}} — Algorithm A PAC learns C with H if, —

Given examples drawn from P, labeled by some c ∈ C: (x1,c(x1)),…,(xn,c(xn))

—

A outputs an α-­‑good hypothesis h ∈ H w.p. 1-β ¡

— Proper: C = H — Fact: Θ(VC(C)) samples for PAC learning C (properly)

¡ VC(C) ≤ log|C|

slide-21
SLIDE 21

PRIVATE LEARNING

slide-22
SLIDE 22

WHY PRIVATE LEARNING?

— Party line [KLNRS’08]: abstracts many of the computations

done over collections of sensitive information

— Test-bed for ideas – problems and mitigation

— Learning intimately related with differential

privacy

¡ Learning theory tools useful for privacy [BLR’08, HR’10, …] ¡ Differential privacy implies generalization [M’?, DFHPRR’15,

BSSU’15, NS’15]

÷ In a sense, all differential privacy allows us is to learn!

slide-23
SLIDE 23

PRIVATE LEARNING

— Definition [KLNRS’08]:

¡ Algorithm A Private-PAC learns C with H if,

÷ A PAC learns C with H, and ÷ A is (ε, δ)–differentially private

slide-24
SLIDE 24

Given labeled examples 𝑻=​(​𝒚↓

𝒚↓𝒋 ,​𝒛↓ 𝒛↓𝒋 )↓𝒋=𝟐↑𝒏 ↑𝒏 provide:

1) Differential Privacy 2) If 𝑻 is consistent with some ​𝒅↓𝒌

𝒅↓𝒌 ∈𝐐𝐏𝐉 𝐏𝐉𝐎​𝐔↓𝒆 ↓𝒆 , then

w.h.p. outputs an 𝒊 s.t.

𝒇𝒔 𝒇𝒔𝒔𝒑​𝒔↓ 𝒔↓𝑻 (𝒊)=​𝟐/𝒏 /𝒏 |{𝒋 ¡:𝒊(​𝒚↓ 𝒚↓𝒋 )≠​𝒛↓ 𝒛↓𝒋 }|≤𝜷

E

𝑸𝑷 𝑸𝑷𝑱𝑶​𝑼↓ 𝑼↓𝒆 ={ ¡█□⁠.⁠ ¡ ¡ ¡𝟏 ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡𝒌 ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡𝑼 ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡} ​𝒅↓𝒌 𝒅↓𝒌 (𝒚)=𝟐 ¡ ¡⟺ ¡ ¡𝒚=𝒌 ​𝒅↓𝒌 𝒅↓𝒌 (𝒚) 𝒚

Example 1: Privately Learning Points

slide-25
SLIDE 25

Example 2: Privately Learning Thresholds

Given labeled examples 𝑻=​(​𝒚↓

𝒚↓𝒋 ,​𝒛↓ 𝒛↓𝒋 )↓𝒋=𝟐↑𝒏 ↑𝒏 provide:

1) Differential Privacy 2) If 𝑻 is consistent with some ​𝒅↓𝒌

𝒅↓𝒌 ∈𝐔𝐈𝐒𝐅𝐓𝐈𝐏𝐌 𝐏𝐌​𝐄↓𝒆 ↓𝒆 ,

then w.h.p. outputs an 𝒊 s.t.

𝒇𝒔 𝒇𝒔𝒔𝒑​𝒔↓ 𝒔↓𝑻 (𝒊)=​𝟐/𝒏 /𝒏 |{𝒋 ¡:𝒊(​𝒚↓ 𝒚↓𝒋 )≠​𝒛↓ 𝒛↓𝒋 }|≤𝜷

𝐔𝐈𝐒𝐅𝐅𝐓𝐈𝐏𝐌 𝐏𝐌​𝐄↓𝒆 ↓𝒆 ={ ¡█□⁠.⁠ ¡ ¡ ¡𝟏 ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡𝒌 ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡𝑼 ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ } ​𝒅↓𝒌 𝒅↓𝒌 (𝒚)=𝟐 ¡ ¡⟺ ¡ ¡𝒚<𝒌 ​𝒅↓𝒌 𝒅↓𝒌 (𝒚) 𝒚

slide-26
SLIDE 26

A General Feasibility Result

— Theorem [KLNRS 08]: Every finite concept class C

can be learned privately (and properly), using O(log| C|) examples

— Generic Construction (based on the Exponential

Mechanism of [MT07]):

¡ Define q(D,h) = # of xi’s correctly classified by h

¡ Output hypothesis h from C w.p. ≈ eεq(D,h)

q(D,h)=3 q(D,h)=4

slide-27
SLIDE 27

A General Feasibility Result

— Generic Construction (based on the Exponential

Mechanism of [MT07]):

¡ Define q(D,h) = # of xi’s incorrectly classified by h

¡ Output hypothesis h from C w.p. ≈ e-εq(D,h)

— Privacy:

¡ changing one example changes q(D,h) by at most 1 ¡ Probability of outputting h changes by a factor of at most eε

— Utility:

¡ If h has error > α, probability of outputting h is at most e-εαn ¡ Union bound: probability of outputting some h with error > α at

most |C|e-εαn

¡ Suffices to take n =O(log |C|)

slide-28
SLIDE 28

Had it been the year 2008 …

slide-29
SLIDE 29

THANK YOU!

Brilliant, isn’t it? Time for coffee and cookies

? ? ?

But …

slide-30
SLIDE 30

Privately Learning Points/Thresholds

— Fact: Proper Point/Threshold learner with O(1)

samples

— Generic construction of private learners results in

O(log |C|) = O(log(T)) samples

— Why do we care?

¡ Want private learners to be as efficient as non-private ones ¡ Generic construction fails when domain infinite

— Thm [BKN 10]: Any proper pure-private learner of

Points/Threshold must use Ω(log​(𝑈)) samples

2d Is this gap essential?

slide-31
SLIDE 31

CAN WE DO BETTER?

— Recall: O(log|C|) examples to beat union bound in

exponential mechanism analysis

— Idea: what if we choose the outcome hypothesis

from a set smaller than C?

slide-32
SLIDE 32

Representation of concept classes [BKN10, BNS11]

— Deterministic Representation for class 𝓓:

¡ DRep: A hypothesis class ℋ s.t.

÷ for every 𝑑∈𝒟 and distribution P over examples, there exists a

hypothesis ℎ∈ℋ s.t. errorP,c(h) ≤ ¼.

÷ The size of DRep is ln |ℋ|

¡ POINTd: DRep = Θ(log log T)

÷ Yields improper learner with sample complexity O(log log T)

— Probabilistic Representation for class 𝓓:

¡ Rep: List of hypothesis classes ℋ1,ℋ2,…,ℋ𝑠 s.t.

÷ for every 𝑑∈𝒟 and distribution P over examples, ÷ w.p. ¾ , a randomly chosen ℋ𝑗 contains a hypothesis ℎ∈ℋ s.t.

errorP,c(h) ≤ ¼.

÷ The size of Rep is defined as maxi (ln |ℋ𝑗|)

— 𝑆𝑓𝑞𝐸𝑗𝑛(𝒟): the size of C’s minimal probabilistic representation

slide-33
SLIDE 33

A Tight Characterization of pure-private learners [BNS 13]

— Define 𝑆𝑓𝑞𝐸𝑗𝑛(𝒟): the size of C’s minimal probabilistic

representation

— Theorem: Θ(𝑆𝑓𝑞𝐸𝑗𝑛(𝒟)) samples are necessary and

sufficient for pure-privately learning 𝒟

¡ Analogous to the VC dimension for non-private learners

thresholds Proper

Θ(​log⁠(𝑈) ) [KLNRS’08, BKN’10]

Improper

Θ(log​(𝑈)) [FX’13]

points Proper

Θ(log​(𝑈)) [KLNRS’08, BKN’10]

Improper

Θ(1) [BKN’10, BNS’13]

Concept class learner Sample complexity C Proper

𝑃(​log⁠|𝐷| ) [KLNRS’08]

Improper

Θ(𝑆𝑓𝑞𝐸𝑗𝑛(𝐷)) [BNS’13]

slide-34
SLIDE 34

M I T I G A T I N G T H E S A M P L E C O M P L E X I T Y O F P R I V A T E L E A R N E R S

slide-35
SLIDE 35

STRATEGIES TO EXPLORE

— Improper learning [BKN’10, BNS’13] — When labeled examples expensive, unlabeled

examples cheap:

¡ Active learning [BF’15, BNS’15] ¡ Semi-supervised learning [BNS’15]

— Approximate privacy [BNS’14, BNSV’15]

slide-36
SLIDE 36

SEMI-SUPERVISED LEARNING [BNS’15]

— Input: batches of labeled and unlabeled samples — Generic construction:

Every finite concept class 𝑫 can be learned privately can be learned privately using 𝑷(𝐖𝐃

𝐖𝐃(𝑫)) labeled examples.

¡ The construction uses O(|C|) unlabeled examples

— Boosting the labeled sample complexity:

Given a private learner for a concept class 𝑫, it is , it is possible to reduce its labeled sample complexity to

𝑷(𝐖𝐃 𝐖𝐃(𝑫)).

¡ While maintaining the unlabeled sample complexity

slide-37
SLIDE 37

REDUCING THE LABELED SAMPLE COMPLEXITY OF A GIVEN

LEARNER 𝓑

— Base learner 𝓑 with sample

complexity 𝒐.

— Input: Database 𝑻 of size 𝒐,

partially labeled

1.

Let 𝑰 be the set of all dichotomies over 𝑻 realized by the target concept class 𝑫

2.

Choose 𝒊∈𝑰 using the exponential mechanism with the labeled portion of 𝑻

3.

Relabel 𝑻 using 𝒊,

4.

Execute 𝓑

​𝒚↓ 𝒚↓𝟐 , ​𝒛↓ 𝒛↓𝟐 ​𝒚↓ 𝒚↓𝟑 , ​𝒛↓ 𝒛↓𝟑 ​𝒚↓ 𝒚↓𝟒 , ​𝒛↓ 𝒛↓𝟒 ​𝒚↓ 𝒚↓𝟓 , ​𝒛↓ 𝒛↓𝟓 ⋮ ​𝒚↓ 𝒚↓𝒖 , ​𝒛↓ 𝒛↓𝒖 ​𝒚↓ 𝒚↓𝒖+𝟐 , ? ​𝒚↓ 𝒚↓𝒖+𝟑 , ? ​𝒚↓ 𝒚↓𝒖+𝟒 , ? ⋮ ​𝒚↓𝒐 𝒐 , ? ​𝒛 𝒛 ↓𝟑 ​𝒛 𝒛 ↓𝒖 +𝟐 ​𝒛 𝒛 ↓𝒖 +𝟑 ​𝒛 𝒛 ↓𝒖 +𝟒 ​𝒛 ↓𝒐 𝒐

slide-38
SLIDE 38

REDUCING THE LABELED SAMPLE COMPLEXITY OF A GIVEN

LEARNER 𝓑

— Base learner 𝓑 with sample

complexity 𝒐.

— Input: Database 𝑻 of size 𝒐,

partially labeled

1.

Let 𝑰 be the set of all dichotomies over 𝑻 realized by the target concept class 𝑫

2.

Choose 𝒊∈𝑰 using the exponential mechanism with the labeled portion of 𝑻

3.

Relabel 𝑻 using 𝒊

4.

Execute 𝓑

∃𝐠∈𝐈 s.t. ​𝐟𝐬𝐬𝐩𝐬 𝐩𝐬↓𝐓 (𝐠)=𝟏 s.t. ​𝐟𝐬𝐬𝐩𝐬 𝐩𝐬↓𝐓 (𝐠)=𝟏 If 𝐓 contains ≈𝐖𝐃 𝐖𝐃(𝐃)​ contains ≈𝐖𝐃 𝐖𝐃(𝐃)​ 𝐦𝐩𝐡 𝐩𝐡⁠|𝑻| labeled exampless then 𝐢 is is close to the target concept 𝓑 returns a hypothesis returns a hypothesis that is close to 𝐢

slide-39
SLIDE 39

REDUCING THE LABELED SAMPLE COMPLEXITY OF A GIVEN

LEARNER 𝓑

— Base learner 𝓑 with sample

complexity 𝒐.

— Input: Database 𝑻 of size 𝒐,

partially labeled

1.

Let 𝑰 be the set of all dichotomies over 𝑻 realized by the target concept class 𝑫

2.

Choose 𝒊∈𝑰 using the exponential mechanism with the labeled portion of 𝑻

3.

Relabel 𝑻 using 𝒊

4.

Execute 𝓑

Difficulty: H depends on S! Outputting h would breach privacy! Solution: use h to relabel sample, analyze distribution of relabeled databases

slide-40
SLIDE 40

BACK TO LEARNING THRESHOLDS

. . . . . .

0 1 2 3 . . . . . . T

2 24 57 83 121 153 176 182 Vitamin C level x x x x

slide-41
SLIDE 41

WHY LEARN THRESHOLDS?

— Seems fundamental and simple, yet disturbing

difference between private and non-private setting

— [BNSV’15] Get four in the price of one:

¡ Distribution learning: ÷ D – unknown distrib over X with cumulative FD ÷ Goal: Given oracle access to D, find F: X à [0,1] with small

|F(x)-FD(x)| for all x ∈ X

¡ Query release: ÷ Given points (x1,…,xn) ∈ X, output data structure approximating

|{i : xi < z}|/n for all z ∈ X

¡ (Approximate) Median: ÷ Given points (x1,…,xn) ∈ X, output z such that (approx.) half the

points are smaller/greater than z

¡ Interior point: ÷ Given points (x1,…,xn) ∈ X, output z between min and max points

slide-42
SLIDE 42

New results [BNSV’15]

LEARNING THRESHOLDS, DISTRIB. LEARNING, QUERY RELEASE, MEDIAN, INTERIOR POINT

— Pure privacy:

¡ Θ (log T) samples [KLNRS’08, BNS’10, FX’13]

— Approx. privacy:

¡ O(8 log*T) samples [BNS’14]

— Non-privately:

¡ O(1) samples

Is this gap essential?

  • O(2 log*T) samples
  • Ω(log*T) samples
slide-43
SLIDE 43

SOLVING INTERIOR POINT WITH PPPROX DP REQUIRES Ω(LOG*T) SAMPLES[BNVS’15]

— Observe: Impossible to have 𝒐=𝟐 when T ≥ 2 — Induction:

¡ Approx. dp mechanism M for solving IP over T(n+1) w/ n+1 samples à Approx. dp mechanism M’ for solving IP over T(n) w/ n samples ¡ Where T(n+1) = b(n)T(n)

𝟐 ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡𝟑 𝟐 ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡ ¡𝟑

Output 1 w.p. ≥​3/4 Output 1 w.p. ≥​3/4 −𝜀/​𝑓↑𝜗 >​

1/4

slide-44
SLIDE 44

M’

​𝑦↓1 ,​𝑦↓2 ,…, ¡​𝑦↓𝑜 ∈[𝑈(𝑜)] ​ 𝑦

SOLVING INTERIOR POINT WITH PPPROX DP REQUIRES Ω(LOG*T) SAMPLES[BNVS’15]

— If M approx private so is M’ — Suppose M succeeds

¡ ​𝑨 ≥​min⁠(​𝑨↓1 ,…, ¡​𝑨↓𝑜 )

¡ Hence, ​𝑦 ≥min​(​𝑦↓1 ,…, ¡​𝑦↓𝑜 )

— Let 𝑥=​max⁠(​𝑦↓1 ,…, ¡​𝑦↓𝑜 )

¡ If ​𝑦 >𝑥 then ​𝑨 reveals ​𝑧↓0↑𝑥

+1

¡ By approx. privacy, this can

happen with probability at most ​

𝑓↑𝜗 /𝑐 +𝜀

​y↓i↑1 ​y↓i↑2 …​𝑧↓𝑗↑𝑈(𝑜) ​∈↓𝑆 ​𝑐↑𝑈(𝑜) ​𝑨↓𝑗 =​𝑧↓0↑1 ​𝑧↓0↑2 …​𝑧↓0↑​𝑦↓𝑗 ​𝑧↓𝑗↑​ 𝑦↓𝑗 +1 …​𝑧↓𝑗↑𝑈(𝑜) ​𝑨↓0 =​𝑧↓0↑1 ​𝑧↓0↑2 …​ 𝑧↓0↑𝑈(𝑜) ​𝑨↓0 ,​𝑨↓1 ,…, ¡​𝑨↓𝑜 ∈[𝑈(𝑜+1)]

M

​𝑨 ​𝑦 =|𝑞𝑠𝑓𝑔𝑗𝑦(​𝑨 , ¡​𝑨↓0 )|

slide-45
SLIDE 45

NOTES ON LOWERBOUND

— Can be stated using a (generalization of)

fingerprinting codes

¡ Fingerprinting codes used to prove other lowerbounds in

differential privacy [BUV14, DTTZ14, BST14]

— Other bounds:

¡ Approx private learning of d-dim threshold funcs over [T]d

is Ω(𝑒​log↑∗ ⁠𝑈 )=Ω(𝑊𝐷​log↑∗ ⁠𝑈 )

¡ Improper pure-privacy learning of POINT over countably

infinite domains

÷ [BNS13]: O(1) samples, but infinite description length

hypotheses

÷ New: This is inherent!

slide-46
SLIDE 46

Private Learning – Related Work

(very partial list)

— [DMNS 05] First private learning algorithms. SQ

based.

— [KLNRS 08] Define private learning, characterize

classification problems learnable privately, understand power of popular models for private analysis.

— [BKN 10] Sample complexity of private learning. — [CH 11] Learning in continuous domain, label privacy. — [CM 08, …] Machine learning. — [BLR 08, DNRRV 09, …] Synthetic Data. — [DRV 10] Private Boosting.

slide-47
SLIDE 47

WHAT HAVE WE LEARNED?

— Private PAC learning exhibits a lot of complexity:

¡ Even for simple complexity classes like points and thresholds

÷ Behaves very differently from non-private learning!

¡ A variety of applicable strategies:

÷ Improper vs. proper for pure DP

¢ RepDim characterizes sample complexity

÷ Semi-supervised

¢ VC labeled samples suffice

÷ Active learning ÷ Label privacy

¢ VC characterizes sample complexity

÷ Approx. vs. pure DP

¢ δ > 0 helps! ¢ Open: improper, approx. privacy ¢ Characterization of sample complexity under approx. privacy

¡ Open: time complexity of private PAC learning

slide-48
SLIDE 48

THANK YOU!

! ! !

Oh, well ! Time for coffee and cookies

slide-49
SLIDE 49

THRESHOLDS AND COMPUTATIONAL COMPLEXITY [BUN-ZHANDRY’15]

— Order Revealing Encryption: — Learn {<, >, =} but nothing else — Efficiently learnable, but not efficiently privately learnable

Thanks: Mark Bun . . . . . .

0 1 2 3 . . . . . . T

2 24 57 83 121 153 176 182 x x x x