privacy and fault tolerance in distributed optimization
play

Privacy and Fault-Tolerance in Distributed Optimization Nitin - PowerPoint PPT Presentation

Privacy and Fault-Tolerance in Distributed Optimization Nitin Vaidya University of Illinois at Urbana-Champaign Acknowledgements Shripad Gade Lili Su S X f i ( x ) argmin x X i =1 i Applications f 1 (x) g fi(x) = cost for robot i


  1. Privacy and Fault-Tolerance in Distributed Optimization Nitin Vaidya University of Illinois at Urbana-Champaign

  2. Acknowledgements Shripad Gade Lili Su

  3. S X f i ( x ) ∈ argmin x ∈ X i =1 i

  4. Applications f 1 (x) g fi(x) = cost for robot i x to go to location x x 1 g Minimize total cost of rendezvous x 2 S f 2 (x) X f i ( x ) ∈ argmin x ∈ X i =1 i

  5. Applications f 1 (x) f 2 (x) Learning Minimize cost Σ f i (x) i f 3 (x) f 4 (x) 5

  6. Outline S X f i ( x ) ∈ argmin x ∈ X i =1 i 𝑔 𝑔 𝑔 𝑔 𝑔 𝑔 & & & " " " 𝑔 𝑔 𝑔 𝑔 𝑔 𝑔 # # # % % % 𝑔 𝑔 𝑔 $ $ $ Distributed Privacy Fault-tolerance Optimization

  7. Distributed Optimization Server 𝑔 𝑔 & " 𝑔 𝑔 # % 𝑔 𝑔 𝑔 𝑔 & # " $ 7

  8. Client-Server Architecture Server f 1 (x) f 2 (x) 𝑔 𝑔 𝑔 & # " 8 f 3 (x) f 4 (x)

  9. Client-Server Architecture g Server maintains estimate 𝑦 ( 𝑦 ( g Client i knows 𝑔 ) (𝑦) Server 𝑔 𝑔 𝑔 & # "

  10. Client-Server Architecture g Server maintains estimate 𝑦 ( 𝑦 ( g Client i knows 𝑔 ) (𝑦) Server 𝛼𝑔 ) (𝑦 ( ) In iteration k+1 g Client i 𝑔 𝑔 𝑔 & # " i Download 𝑦 ( from server i Upload gradient 𝛼𝑔 ) (𝑦 ( )

  11. � Client-Server Architecture g Server maintains estimate 𝑦 ( g Client i knows 𝑔 ) (𝑦) Server 𝛼𝑔 ) (𝑦 ( ) In iteration k+1 g Client i 𝑔 𝑔 𝑔 & # " i Download 𝑦 ( from server i Upload gradient 𝛼𝑔 ) (𝑦 ( ) g Server 𝑦 (-& ⟵ 𝑦 ( − 𝛽 ( 2 𝛼𝑔 ) 𝑦 ( )

  12. Variations g Stochastic g Asynchronous g … 12

  13. Peer-to-Peer Architecture f 1 (x) f 2 (x) 𝑔 𝑔 & " 𝑔 𝑔 # % 𝑔 $ f 3 (x) f 4 (x)

  14. Peer-to-Peer Architecture g Each agent maintains local estimate x g Consensus step with neighbors g Apply own gradient to own estimate 𝑦 (-& ⟵ 𝑦 ( − 𝛽 ( 𝛼𝑔 ) 𝑦 ( 𝑔 𝑔 & " 𝑔 𝑔 # % 𝑔 $

  15. Outline S X f i ( x ) ∈ argmin x ∈ X i =1 i 𝑔 𝑔 𝑔 𝑔 𝑔 𝑔 & & & " " " 𝑔 𝑔 𝑔 𝑔 𝑔 𝑔 # # # % % % 𝑔 𝑔 𝑔 $ $ $ Distributed Privacy Fault-tolerance Optimization

  16. Server 𝛼𝑔 ) (𝑦 ( ) 𝑔 𝑔 𝑔 & # "

  17. Server 𝛼𝑔 ) (𝑦 ( ) 𝑔 𝑔 𝑔 & # " Server observes gradients è privacy compromised

  18. Server 𝛼𝑔 ) (𝑦 ( ) 𝑔 𝑔 𝑔 & # " Server observes gradients è privacy compromised Achieve privacy and yet collaboratively optimize

  19. Related Work g Cryptographic methods (homomorphic encryption) g Function transformation g Differential privacy 19

  20. Differential Privacy Server 𝛼𝑔 ) 𝑦 ( + 𝜻 𝒍 𝑔 𝑔 𝑔 & # " 20

  21. Differential Privacy Server 𝛼𝑔 ) 𝑦 ( + 𝜻 𝒍 𝑔 𝑔 𝑔 & # " Trade-off privacy with accuracy 21

  22. Proposed Approach g Motivated by secret sharing g Exploit diversity … Multiple servers / neighbors 22

  23. Proposed Approach Server 2 Server 1 𝑔 𝑔 𝑔 & # " Privacy if subset of servers adversarial 23

  24. Proposed Approach 𝑔 𝑔 & " 𝑔 𝑔 # % 𝑔 $ Privacy if subset of neighbors adversarial 24

  25. Proposed Approach g Structured noise that “cancels” over servers/neighbors 25

  26. Intuition x 2 x 1 Server 2 Server 1 𝑔 𝑔 𝑔 " & # 26

  27. Intuition x 2 x 1 Server 2 Server 1 𝑔 𝑔 𝑔 #& 𝑔 𝑔 "& 𝑔 && &# ## "# Each client simulates multiple clients 27

  28. Intuition x 2 x 1 Server 2 Server 1 𝑔 𝑔 𝑔 #& 𝑔 𝑔 "& 𝑔 && &# ## "# 𝑔 && (𝑦) + 𝑔 &# 𝑦 = 𝑔 & 𝑦 𝑔 )8 (𝑦) not necessarily convex 28

  29. Algorithm g Each server maintains an estimate In each iteration g Client i i Download estimates from corresponding server i Upload gradient of 𝑔 ) g Each server updates estimate using received gradients

  30. Algorithm g Each server maintains an estimate In each iteration g Client i i Download estimates from corresponding server i Upload gradient of 𝑔 ) g Each server updates estimate using received gradients g Servers periodically exchange estimates to perform a consensus step

  31. Claim g Under suitable assumptions, servers eventually reach consensus in S X f i ( x ) ∈ argmin x ∈ X i =1 i 31

  32. Privacy 𝑔 && + 𝑔 #& + 𝑔 𝑔 #& + 𝑔 ## + 𝑔 "& "# Server 2 Server 1 𝑔 𝑔 𝑔 #& 𝑔 𝑔 "& 𝑔 && &# ## "# 32

  33. Privacy 𝑔 && + 𝑔 #& + 𝑔 𝑔 #& + 𝑔 ## + 𝑔 "& "# Server 2 Server 1 𝑔 𝑔 𝑔 #& 𝑔 𝑔 "& 𝑔 && &# ## "# g Server 1 may learn 𝑔 && , 𝑔 #& , 𝑔 "& , 𝑔 #& + 𝑔 ## + 𝑔 "# g Not sufficient to learn 𝑔 ) 33

  34. 𝑔 && (𝑦) + 𝑔 &# 𝑦 = 𝑔 & 𝑦 g Function splitting not necessarily practical g Structured randomization as an alternative 34

  35. Structured Randomization g Multiplicative or additive noise in gradients g Noise cancels over servers 35

  36. Multiplicative Noise x 2 x 1 Server 2 Server 1 𝑔 𝑔 𝑔 " & # 36

  37. Multiplicative Noise x 2 x 1 Server 2 Server 1 𝑔 𝑔 𝑔 " & # 37

  38. Multiplicative Noise x 2 x 1 Server 2 Server 1 𝛽𝛼𝑔 𝛾𝛼𝑔 & (𝑦 2 ) & (x 1 ) 𝑔 𝑔 𝑔 " & # 𝛽 + 𝛾 =1 38

  39. Multiplicative Noise x 2 x 1 Server 2 Server 1 𝛽𝛼𝑔 𝛾𝛼𝑔 & (𝑦 2 ) & (x 1 ) 𝑔 𝑔 𝑔 " & # Suffices for this invariant to hold 𝛽 + 𝛾 =1 over a larger number of iterations

  40. Multiplicative Noise x 2 x 1 Server 2 Server 1 𝛽𝛼𝑔 𝛾𝛼𝑔 & (𝑦 2 ) & (x 1 ) 𝑔 𝑔 𝑔 " & # Noise from client i to server j 𝛽 + 𝛾 =1 not zero-mean

  41. Claim g Under suitable assumptions, servers eventually reach consensus in S X f i ( x ) ∈ argmin x ∈ X i =1 i 41

  42. Peer-to-Peer Architecture 𝑔 𝑔 & " 𝑔 𝑔 # % 𝑔 $

  43. Reminder … g Each agent maintains local estimate x g Consensus step with neighbors g Apply own gradient to own estimate 𝑦 (-& ⟵ 𝑦 ( − 𝛽 ( 𝛼𝑔 ) 𝑦 ( 𝑔 𝑔 & " 𝑔 𝑔 # % 𝑔 $

  44. Proposed Approach g Each agent shares noisy estimate with neighbors • Scheme 1 – Noise cancels over neighbors • Scheme 2 – Noise cancels network-wide 𝑔 𝑔 & " 𝑔 𝑔 # % 𝑔 $

  45. Proposed Approach g Each agent shares noisy estimate with neighbors • Scheme 1 – Noise cancels over neighbors • Scheme 2 – Noise cancels network-wide x + ε 1 ε 1 + ε 2 = 0 (over iterations) 𝑔 𝑔 & " x + ε 2 𝑔 𝑔 # % 𝑔 $

  46. Peer-to-Peer Architecture g Poster today Shripad Gade

  47. Outline S X f i ( x ) ∈ argmin x ∈ X i =1 i 𝑔 𝑔 𝑔 𝑔 𝑔 𝑔 & & & " " " 𝑔 𝑔 𝑔 𝑔 𝑔 𝑔 # # # % % % 𝑔 𝑔 𝑔 $ $ $ Distributed Privacy Fault-tolerance Optimization

  48. Fault-Tolerance g Some agents may be faulty g Need to produce “correct” output despite the faults 48

  49. Byzantine Fault Model g No constraint on misbehavior of a faulty agent g May send bogus messages g Faulty agents can collude 49

  50. Peer-to-Peer Architecture f 1 (x) g fi(x) = cost for robot i x to go to location x x 1 g Faulty agent may choose arbitrary cost function x 2 f 2 (x)

  51. Peer-to-Peer Architecture 𝑔 𝑔 & " 𝑔 𝑔 # % 𝑔 $ 51

  52. Client-Server Architecture Server 𝛼𝑔 ) (𝑦 ( ) 𝑔 𝑔 𝑔 & # "

  53. Fault-Tolerant Optimization g The original problem is not meaningful S X f i ( x ) ∈ argmin x ∈ X i =1 i 53

  54. Fault-Tolerant Optimization g The original problem is not meaningful S X f i ( x ) ∈ argmin x ∈ X i =1 i g Optimize cost over only non-faulty agents S X f i ( x ) ∈ argmin x ∈ X i =1 i good

  55. Fault-Tolerant Optimization g The original problem is not meaningful S X f i ( x ) ∈ argmin x ∈ X i =1 i g Optimize cost over only non-faulty agents S X f i ( x ) ∈ argmin Impossible! x ∈ X i =1 i good

  56. Fault-Tolerant Optimization g Optimize weighted cost over only non-faulty agents S X 𝛃 i f i ( x ) ∈ argmin x ∈ X i =1 i good g With 𝛃 i as close to 1/ good as possible

  57. Fault-Tolerant Optimization g Optimize weighted cost over only non-faulty agents S X 𝛃 i f i ( x ) ∈ argmin x ∈ X i =1 i good With t Byzantine faulty agents: t weights may be 0

  58. Fault-Tolerant Optimization g Optimize weighted cost over only non-faulty agents S X 𝛃 i f i ( x ) ∈ argmin x ∈ X i =1 i good t Byzantine agents, n total agents At least n-2t weights guaranteed to be > 1/2(n-t)

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend