split learning
play

Split Learning A resource efficient distributed deep learning - PowerPoint PPT Presentation

Split Learning A resource efficient distributed deep learning method without sensitive data sharing Praneeth Vepakomma vepakom@mit.edu Invisible Health Image Data Small Data Small Data Small Data ML for Health


  1. Split Learning A resource efficient distributed deep learning method without sensitive data sharing Praneeth Vepakomma vepakom@mit.edu

  2. ‘Invisible’ Health Image Data ‘Small Data’ ‘Small Data’ ‘Small Data’

  3. ML for Health Images Low a. Distributed Data Bandwidth b. Patient privacy c. Incentives d. ML Expertise ‘Small’ Data e. Efficiency Low Compute

  4. Train Neural Nets No Exchange of Raw Patient Images Gupta, Raskar ‘Distributed training of deep neural network over several agents’, 2017

  5. Intelligent Computing Security, Privacy & Safety

  6. Regulations GDPR: General Data Protection Regulation HIPAA: Health Insurance Portability and Accountability Act, 1996 SOX : Sarbanes-Oxley Act, 2002 PCI : Payment Card Industry Data Security Standard, 2004 SHIELD: Stop Hacks and Improve Electronic Data Security Act, Jan 1 2019

  7. Challenges for Distributed Data + AI + Health Distributed Data Regulations Multi-Modal Incentives Incomplete Data Cooperation Ease Ledgering Resource-constraints Smart contracts Memory, Compute, Bandwidth, Maintenance Convergence, Synchronization, Leakage

  8. AI: Bringing it all together Training No sharing of Deep Raw Images Networks Server Client Invisible Data / Data Friction

  9. Overcoming Data Friction Ease Incentive Trust Regulation Blockchain AI/ SplitNN

  10. Anonymize Obfuscate Encrypt Protect Data

  11. Data Utility Train Model Share Wisdom Infer Statistics Data Protect Anonymize Obfuscate Smash Encrypt Hide Raw Add Noise Private Data

  12. Federated Learning Nets trained at Clients Merged at Server Differential Privacy Split Learning (MIT) Obfuscate with noise Nets split over network Hide unique samples Trained at both Homomorphic Encryption Basic Math over Encrypted Data (+, x)

  13. Federated Learning Server Client1 Client2 Client3 ..

  14. Protect Partial Differential Homomorphic data Oblivious Transfer, Garbled Distributed Leakage Circuits Privacy Encryption Training Federated Learning Split Learning Inference but no training Praneeth Vepakomma, Tristan Swedish, Otkrist Gupta, Abhi Dubey, Raskar 2018

  15. When to use split learning? Large number of clients: Split learning shows positive results Split Memory Compute Bandwidth Federated Convergence Project Page and Papers: https://splitlearning.github.io/

  16. Label Sharing No Label Sharing

  17. Gupta, Otkrist, and Raskar, Ramesh. "Secure Training of Multi-Party Deep Neural Network." U.S. Patent Application No. 15/630,944.

  18. Distribution of parameters in AlexNet

  19. Versatile Configurations of Split Learning Split learning for health: Distributed deep learning without sharing raw patient data, Praneeth Vepakomma, Otkrist Gupta, Tristan Swedish, Ramesh Raskar, (2019)

  20. NoPeek SplitNN: Reducing Leakage in Distributed Deep Learning Reducing leakage in distributed deep learning for sensitive health data, Praneeth Vepakomma, Otkrist Gupta, Abhimanyu Dubey, Ramesh Raskar (2019)

  21. No peak deep learning with conditioning variable Setup: Ideal Goal: To find such a conditioning variable Z within the framework of deep learning such that the following directions are approximately satisfied: 1. Y X | Z (Utility property as X can be thrown away given Z to obtain prediction E(Y|Z)) 2. X Z (One-way property preventing proper reconstruction of raw data X from Z) Note: denotes statistical independence

  22. Possible measures of non-linear dependence ● COCO: Constrained Covariance ● HSIC: Hilbert-Schmidt Independence Criterion ● DCOR: Distance Correlation ● MMD: Maximum Mean Discrepancy ● KTA: Kernel Target Alignment ● MIC: Maximal Information Coefficient ● TIC: Total Information Coefficient

  23. Why is it called distance correlation?

  24. Praneeth Vepakomma, Chetan Tonde, Ahmed Elgammal, Electronic Journal of Statistics, 2018

  25. Colorectal histology image dataset (Public data)

  26. Leakage Reduction in Action Reduced leakage during training Reduced leakage during training over colorectal histology image over colorectal histology image data from 0.96 in traditional CNN to data from 0.92 in traditional CNN to 0.19 in NoPeek SplitNN 0.33 in NoPeek SplitNN Reducing leakage in distributed deep learning for sensitive health data, Praneeth Vepakomma, Otkrist Gupta, Abhimanyu Dubey, Ramesh Raskar (2019)

  27. Similar validation performance

  28. Effect of leakage reduction on convergence

  29. Robustness to reconstruction

  30. Proof of one-Way Property: We show: Minimizing regularized distance covariance minimizes the difference of Kullback-Leibler divergences

  31. Project Page and Papers: https://splitlearning.github.io/ Thanks and acknowledgements to: Otkrist Gupta (MIT/LendBuzz), Ramesh Raskar (MIT), Jayashree Kalpathy-Cramer (Martinos/Harvard), Rajiv Gupta (MGH), Brendan McMahan (Google), Jakub Kone č ný (Google), Abhimanyu Dubey (MIT), Tristan Swedish (MIT), Sai Sri Sathya (S20.ai) , Vitor Pamplona (MIT/EyeNetra), Rodmy Paredes Alfaro (MIT), Kevin Pho (MIT), Elsa Itambo (MIT)

  32. THANK YOU

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend