Graph Analysis with Node Differential Privacy Node Differential - - PowerPoint PPT Presentation

graph analysis with node differential privacy node
SMART_READER_LITE
LIVE PREVIEW

Graph Analysis with Node Differential Privacy Node Differential - - PowerPoint PPT Presentation

Graph Analysis with Node Differential Privacy Node Differential Privacy Sofya Sofya Raskhodnikova Sofya Sofya Raskhodnikova Raskhodnikova Raskhodnikova Penn State University Shiva Kasiviswanathan Shiva Kasiviswanathan (GE Research),


slide-1
SLIDE 1

Graph Analysis with Node Differential Privacy Node Differential Privacy

Sofya Sofya Raskhodnikova Raskhodnikova Sofya Sofya Raskhodnikova Raskhodnikova

Penn State University

Joint work with Shiva Shiva Kasiviswanathan Kasiviswanathan (GE Research), Kobbi Kobbi Nissim Nissim (Ben-Gurion U. and Harvard U.), Ad S i h Ad S i h (P S ) Adam Smith Adam Smith (Penn State)

1

slide-2
SLIDE 2

Publishing information about graphs

Many datasets can be represented as graphs

  • “Friendships” in online social network
  • “Friendships” in online social network
  • Financial transactions

E il i ti

  • Email communication
  • Romantic relationships

image source http://community.expressor- software.com/blogs/mtarallo/36-extracting-data- facebook-social-graph-expressor-tutorial.html

Privacy is a big issue!

2

American J. Sociology, Bearman, Moody, Stovel

big issue!

slide-3
SLIDE 3

Private analysis of graph data

Graph G

i

) (

Government, researchers

Trusted curator Users

queries answers )

(

researchers, businesses (or) malicious malicious adversary

T fli i l ili d i

  • Two conflicting goals: utility and privacy

– utility: accurate answers – privacy: ?

3

image source http://www.queticointernetmarketing.com/new-amazing-facebook-photo-mapper/

slide-4
SLIDE 4

Differential privacy for graph data

Graph G

i

) (

Government, researchers

Trusted curator Users

A

queries answers )

(

researchers, businesses (or) malicious malicious adversary

  • Intuition: neighbors are datasets that differ only in some

Intuition: neighbors are datasets that differ only in some information we’d like to hide (e.g., one person’s data)

4

image source http://www.queticointernetmarketing.com/new-amazing-facebook-photo-mapper/

slide-5
SLIDE 5

Two variants of differential privacy for graphs

  • Edge differential privacy

G: Two graphs are neighbors if they differ in one edge.

  • Node differential privacy

G: Two graphs are neighbors if one can be obtained from the other by deleting a node and its adjacent edges by deleting a node and its adjacent edges.

5

slide-6
SLIDE 6

Node differentially private analysis of graphs

Graph G

i

) (

Government, researchers

Trusted curator Users

A

queries answers )

(

researchers, businesses (or) malicious

T fli ti l tilit d i

malicious adversary

  • Two conflicting goals: utility and privacy

– Impossible to get both in the worst case

  • Previously: no node differentially private
  • Previously: no node differentially private

algorithms that are accurate on realistic graphs

6

image source http://www.queticointernetmarketing.com/new-amazing-facebook-photo-mapper/

slide-7
SLIDE 7

Our contributions

  • First node differentially private algorithms that are

accurate for sparse graphs accurate for sparse graphs

– node differentially private for all graphs acc rate for a s bclass f h hi h i l d – accurate for a subclass of graphs, which includes

  • graphs with sublinear (not necessarily constant) degree bound
  • graphs where the tail of the degree distribution is not too heavy

graphs where the tail of the degree distribution is not too heavy

  • dense graphs
  • Techniques for node differentially private algorithms

ec ques o

  • de d

e e a y p a e a go s

  • Methodology for analyzing the accuracy of such

algorithms on realistic networks algorithms on realistic networks

Concurrent work on node privacy [Blocki Blum Datta Sheffet 13] Concurrent work on node privacy [Blocki Blum Datta Sheffet 13]

7

slide-8
SLIDE 8

Our contributions: algorithms … …

Frequency

Degrees

8

slide-9
SLIDE 9

Our contributions: accuracy analysis

(1+o(1))-approximation

9

slide-10
SLIDE 10

Previous work on

differentially private computations on graphs

Edge differentially private algorithms Edge differentially private algorithms

  • number of triangles, MST cost [Nissim Raskhodnikova Smith 07]
  • degree distribution [Hay Rastogi Miklau Suciu 09 Hay Li Miklau Jensen 09]
  • degree distribution [Hay Rastogi Miklau Suciu 09, Hay Li Miklau Jensen 09]
  • small subgraph counts [Karwa Raskhodnikova Smith Yaroslavtsev 11]
  • cuts [Blocki Blum Datta Sheffet 12]

Edge private against Bayesian adversary (weaker privacy)

ll b h t

[ kl ]

  • small subgraph counts [Rastogi Hay Miklau Suciu 09]

Node zero‐knowledge private (stronger privacy) Node zero knowledge private (stronger privacy)

  • average degree, distances to nearest connected, Eulerian,

cycle‐free graphs for dense graphs [Gehrke Lui Pass 12]

10

slide-11
SLIDE 11

Differential privacy basics

Graph G

t ti ti f )

(

Government, researchers

Trusted curator Users

A

statistic f approximation )

(

researchers, businesses (or) malicious to f(G) malicious adversary

11

slide-12
SLIDE 12

Global sensitivity framework [DMNS’06]

12

slide-13
SLIDE 13

“Projections” on graphs of small degree

Goal: privacy for all graphs

13

slide-14
SLIDE 14

Method 1: Lipschitz extensions

14

slide-15
SLIDE 15

1

1' 1

s 1

3 3'

t

2 2' 1 5 5' 4 4' 15

slide-16
SLIDE 16

1

1' 1 1/

s 1

3 3'

t

2 2' 1 1/ 5 5' 4 4' 16

slide-17
SLIDE 17

1

1' 1

s 1

3 3'

t

2 2' 1 5 5' 4 4' 6' 6 6' 6 17

slide-18
SLIDE 18
  • via Smooth Sensitivity framework [NRS’07]
  • via Smooth Sensitivity framework [NRS 07]

18

slide-19
SLIDE 19

Our results

via Lipschitz extensions

} via generic reduction

19

slide-20
SLIDE 20

Conclusions

  • It is possible to design node differentially private algorithms

with good utility on sparse graphs with good utility on sparse graphs

– One can first test whether the graph is sparse privately

  • Directions for future work

– Node‐private synthetic graphs – What are the right notions of privacy for network data?

20

slide-21
SLIDE 21

Lipschitz extensions via linear/convex programs

21

slide-22
SLIDE 22

Our results … …

Frequency

… …

Degrees

22

slide-23
SLIDE 23

Our results … …

(1+o(1))-approximation (1 o(1)) approximation

23

slide-24
SLIDE 24

T

query f

T(G) G A

24

S

slide-25
SLIDE 25

Generic Reduction via Truncation

Frequency

… …

d Degrees

25

slide-26
SLIDE 26

Smooth Sensitivity of Truncation

#(nodes of degree above d)

26

slide-27
SLIDE 27

Releasing Degree Distribution via Generic Reduction

T

query f

T(G) G A T

q y f

T(G) S

27