Split Learning A resource efficient distributed deep learning - - PowerPoint PPT Presentation

split learning
SMART_READER_LITE
LIVE PREVIEW

Split Learning A resource efficient distributed deep learning - - PowerPoint PPT Presentation

Split Learning A resource efficient distributed deep learning method without sensitive data sharing Praneeth Vepakomma vepakom@mit.edu Invisible Health Image Data Small Data Small Data Small Data ML for Health


slide-1
SLIDE 1

Split Learning

A resource efficient distributed deep learning method without sensitive data sharing Praneeth Vepakomma vepakom@mit.edu

slide-2
SLIDE 2

‘Invisible’ Health Image Data

‘Small Data’ ‘Small Data’ ‘Small Data’

slide-3
SLIDE 3
  • a. Distributed Data
  • b. Patient privacy
  • c. Incentives
  • d. ML Expertise
  • e. Efficiency

Low Bandwidth Low Compute ‘Small’ Data

ML for Health Images

slide-4
SLIDE 4

Gupta, Raskar ‘Distributed training of deep neural network over several agents’, 2017

No Exchange

  • f Raw

Patient Images Train Neural Nets

slide-5
SLIDE 5

Intelligent Computing Security, Privacy & Safety

slide-6
SLIDE 6

GDPR: General Data Protection Regulation HIPAA: Health Insurance Portability and Accountability Act, 1996 SOX: Sarbanes-Oxley Act, 2002 PCI: Payment Card Industry Data Security Standard, 2004 SHIELD: Stop Hacks and Improve Electronic Data Security Act, Jan 1 2019

Regulations

slide-7
SLIDE 7
slide-8
SLIDE 8
slide-9
SLIDE 9

Distributed Data Multi-Modal Incomplete Data Resource-constraints Memory, Compute, Bandwidth, Convergence, Synchronization, Leakage Regulations Incentives Cooperation Ease Ledgering Smart contracts Maintenance

Challenges for Distributed Data + AI + Health

slide-10
SLIDE 10
slide-11
SLIDE 11

Training Deep Networks No sharing of Raw Images

Server Client Invisible Data / Data Friction

AI: Bringing it all together

slide-12
SLIDE 12

Ease Incentive Trust Regulation

Blockchain AI/ SplitNN

Overcoming Data Friction

slide-13
SLIDE 13

Anonymize

Protect Data

Obfuscate Encrypt

slide-14
SLIDE 14
slide-15
SLIDE 15

Hide Raw Data

Share Wisdom

Data Utility

Train Model

Encrypt Smash Obfuscate

Add Noise Private

Data Protect

Infer Statistics

Anonymize

slide-16
SLIDE 16

Federated Learning Nets trained at Clients Merged at Server Differential Privacy Obfuscate with noise Hide unique samples Homomorphic Encryption Basic Math over Encrypted Data (+, x) Split Learning (MIT) Nets split over network Trained at both

slide-17
SLIDE 17

Federated Learning

Server Client1 Client2 Client3 ..

slide-18
SLIDE 18

Partial Leakage Differential Privacy Homomorphic Encryption

Oblivious Transfer, Garbled Circuits Federated Learning Split Learning

Protect data Distributed Training

Inference but no training

Praneeth Vepakomma, Tristan Swedish, Otkrist Gupta, Abhi Dubey, Raskar 2018

slide-19
SLIDE 19

Memory Compute Bandwidth Convergence Federated Split

When to use split learning?

Large number of clients: Split learning shows positive results

Project Page and Papers: https://splitlearning.github.io/

slide-20
SLIDE 20
slide-21
SLIDE 21

Label Sharing No Label Sharing

slide-22
SLIDE 22

Gupta, Otkrist, and Raskar, Ramesh. "Secure Training of Multi-Party Deep Neural Network." U.S. Patent Application No. 15/630,944.

slide-23
SLIDE 23

Distribution of parameters in AlexNet

slide-24
SLIDE 24

Versatile Configurations of Split Learning

Split learning for health: Distributed deep learning without sharing raw patient data, Praneeth Vepakomma, Otkrist Gupta, Tristan Swedish, Ramesh Raskar, (2019)

slide-25
SLIDE 25

Reducing leakage in distributed deep learning for sensitive health data, Praneeth Vepakomma, Otkrist Gupta, Abhimanyu Dubey, Ramesh Raskar (2019)

NoPeek SplitNN: Reducing Leakage in Distributed Deep Learning

slide-26
SLIDE 26

No peak deep learning with conditioning variable Setup:

Ideal Goal: To find such a conditioning variable Z within the framework of deep learning such that the following directions are approximately satisfied: 1. Y X | Z (Utility property as X can be thrown away given Z to obtain prediction E(Y|Z)) 2. X Z (One-way property preventing proper reconstruction of raw data X from Z) Note: denotes statistical independence

slide-27
SLIDE 27
  • COCO: Constrained Covariance
  • HSIC: Hilbert-Schmidt Independence Criterion
  • DCOR: Distance Correlation
  • MMD: Maximum Mean Discrepancy
  • KTA: Kernel Target Alignment
  • MIC: Maximal Information Coefficient
  • TIC: Total Information Coefficient

Possible measures of non-linear dependence

slide-28
SLIDE 28

Why is it called distance correlation?

slide-29
SLIDE 29

Praneeth Vepakomma, Chetan Tonde, Ahmed Elgammal, Electronic Journal of Statistics, 2018

slide-30
SLIDE 30

Colorectal histology image dataset (Public data)

slide-31
SLIDE 31

Leakage Reduction in Action

Reduced leakage during training

  • ver colorectal histology image

data from 0.96 in traditional CNN to 0.19 in NoPeek SplitNN Reduced leakage during training

  • ver colorectal histology image

data from 0.92 in traditional CNN to 0.33 in NoPeek SplitNN

Reducing leakage in distributed deep learning for sensitive health data, Praneeth Vepakomma, Otkrist Gupta, Abhimanyu Dubey, Ramesh Raskar (2019)

slide-32
SLIDE 32

Similar validation performance

slide-33
SLIDE 33

Effect of leakage reduction on convergence

slide-34
SLIDE 34

Robustness to reconstruction

slide-35
SLIDE 35

Proof of one-Way Property: We show: Minimizing regularized distance covariance minimizes the difference of Kullback-Leibler divergences

slide-36
SLIDE 36
slide-37
SLIDE 37
slide-38
SLIDE 38

Project Page and Papers: https://splitlearning.github.io/

Thanks and acknowledgements to: Otkrist Gupta (MIT/LendBuzz), Ramesh Raskar (MIT), Jayashree Kalpathy-Cramer (Martinos/Harvard), Rajiv Gupta (MGH), Brendan McMahan (Google), Jakub Konečný (Google), Abhimanyu Dubey (MIT), Tristan Swedish (MIT), Sai Sri Sathya (S20.ai), Vitor Pamplona (MIT/EyeNetra), Rodmy Paredes Alfaro (MIT), Kevin Pho (MIT), Elsa Itambo (MIT)

slide-39
SLIDE 39

THANK

YOU