CSC2412: Exponential Mechanism & Private PAC Learning
Sasho Nikolov
1
CSC2412: Exponential Mechanism & Private PAC Learning Sasho - - PowerPoint PPT Presentation
CSC2412: Exponential Mechanism & Private PAC Learning Sasho Nikolov 1 Classification Basics The learning problem - I + I . Problem: develop an algorithm that classifies avocados into ripe and unripe. We have a big data set of avocado data.
CSC2412: Exponential Mechanism & Private PAC Learning
Sasho Nikolov
1
Classification Basics
The learning problem
Problem: develop an algorithm that classifies avocados into ripe and unripe. We have a big data set of avocado data. For each avocado, we have:
From this data, we want to classify unseen avocados.
2 + I
features label
Learn
a
rule
to
predict
label
from
features
The learning problem, formally
Model:
independent sample from D. Goal: Learn c from X.
3
all possible setting
to the features
All allow
to my
:
features
to label
.'t done
features
some
cec
can
,
produce
all effete
features -
E
an
avocado
is
ripe
approximation of
e
/ iff
colour
= Attu
and
firmness
= medium
The goal, formally
The error of a concept c0 2 C is LD,c = Px⇠D(c0(x) 6= c(x)). We want an algorithm M that outputs some c0 2 C and satisfies P(LD,c(M(X)) α) 1 β.
4
Fraction
the
population
labeled
↳HWA
incorrectly by
c
'Floss of
c
' ( w . r. t .D
,c )
X
sampled
iid
from
D
^
→ output of UK)
misclassifies
Ed fraction
% taken
randomness
in choosing
X and any
randomness
Probably Approximately
Correct
learning
( PAC)
[ Valiant]
Empirical risk minimization
Issue: We want to find arg minc02C LD,c(c0), but we do not know D, c. Solution: Instead we solve arg minc02C LX(c0), where LX(c0) = |{i : c0(xi) 6= c(xi)}| n is the empirical error. Theorem (Uniform convergence) Suppose that n ln(|C|/β)
2α2
. Then, with probability 1 β, max
c02C LX(c0) LD,c(c0) α. 5
pnpnoxiuate minimizerok LD ,e Population ↳c4=O cuukuol
:L ,
I
fraction of
pts
in X
Lx empirical
loss misclassified by
c
'ckuowil
→ Pop and emp
. lossare
close
for
Kc
'c- C
uoeetoung
'EE1Yt
Other
versions
for infinite
C
, e.g ,VC
Private learning
In private PAC learning, we require that
stadard PAC learning;
X 2 (X ⇥ C)n.
6
ele C
Privacy
must
even
if data X
'
neighbouring
t
s
'
not
iid
e- S )
t
ee
. PINK' ) c- S )
do
we
e
( approximately )
( c ' )
=
I
i
:
Citi )
#
c' exit } I
c
'
C- C
'How
can
we
use
noise
?
Exercise
: analyze
is
a counting query
this f
answers
to
all
,
. . , eh }, 4h27 ,
. - .. ↳ 141 }
Exponential mechanism
Private ERM
We want to solve arg minc02C LX(c0). How do we minimize with differential privacy? Sample concepts with less error with higher probability P(M(X) = c0) / exp ⇣ εn 2 LX(c0) ⌘
7
D
Exponential Mechanism
General set-up: score function u : X ⇥ Y ! R Sensitivity ∆u = max
y2Y max X⇠X 0 |u(X, y) u(X 0, y)|.
The mechanism Mexp(X) which outputs a random Y so that P(Y = y) = eεu(X,y)/2∆u P
z2Y eεu(X,z)/2∆u
is ε-differentially private
8
u (X
, y)
=
" how good of
Ban
an
is y
Goal
: given
X
, find
argq.mg
UH
, y )
for
the dataset
X ?
"
How
different
can
the
score
be between neighbouring
X
, X
'→ normalizing
factor
Privacy analysis
9
Enough
to show
: H X - t
'
PINCH
. )
'
e '
, .e*eu"¥:"#t¥÷::÷÷:
E eEK.EE/Z--eE
Accuracy of the exponential mechanism
OPT(X) = max
y2Y u(X, y)
Then, for the output Y = Mexp(X), P(u(X, Y ) OPT(X) t)
10
ZL
y*
achieves
)
)
a- OPT
z
ee 'd
thou
s
eel
OPT
⇐ e- ethos ?
s
e
peg )
191
Private Learning
Unknown
distribution
D
known R
Unknown
c
in
a
known
concept
class
C
Data
set
,
. - , Hm, um }
where
is
,
. . ., Xu
, etc
' )
( cent
c' Ch ) / ↳ ( c
. ) . . Ki:cHniHdHi#n
z
then
we prob z
t - p,
C
! Lo
, ,
Cc
' )
e- ↳ kilt d
:
we prob
.
to expleulkylb.su)
Au
'
. g) I
Putting things together
A concept class C can be learned by an eps-differentially private mechanism when the sample size is n max ⇢4 ln(2|C|/β) εα , 2 ln(2|C|/β) α2
EEE
Use
exp
.mechanism
with
B
with ult.ci/---LxT-
With
. prob?
I -
R2
,
c
'
by
heap (x)
has
↳ ( C
' ) Etz . Byunit . convergence , we prob z
t - Bf , LD
, etc
' )
Putting things together
12
MIX
, c.)
=
;
Du
= th ;
sample
c' e
C at prof
prop
.to etpl
Optfx)
c' c- C
=
Lx ( c ')
C ' c- C
IPI ↳ ( MH )) z E)
= plulx , UK)) EOPTKI
E
e
if
u z 4lntp)
Ed
+ by YwYfg
w/ prob
I
↳
. . ( UCH) thrill CH) -172upset
? ' -B drink =L
.