submanifold sparse convolutional networks for sparse
play

Submanifold Sparse Convolutional Networks for Sparse, Locally Dense - PowerPoint PPT Presentation

Submanifold Sparse Convolutional Networks for Sparse, Locally Dense Particle Image Analysis Laura Domine (Stanford / SLAC) Kazuhiro Terao (SLAC) 2018 CPAD Instrumentation Frontier Workshop 1 Outline 1. Particle image analysis &


  1. Submanifold Sparse Convolutional Networks for Sparse, Locally Dense Particle Image Analysis Laura Domine (Stanford / SLAC) Kazuhiro Terao (SLAC) 2018 CPAD Instrumentation Frontier Workshop 1

  2. Outline 1. Particle image analysis & Convolutional networks 2. Submanifold Sparse Convolutions 3. Comparison study between a dense and sparse network 2 2018 CPAD / L.Domine and K.Terao

  3. Outline 1. Particle image analysis & Convolutional networks 2. Submanifold Sparse Convolutions 3. Comparison study between a dense and sparse network e 3 2018 CPAD / L.Domine and K.Terao

  4. Particle Image Analysis with LArTPCs Liquid Argon Time Projection Chamber (LArTPC) = particle imaging detector ~3mm resolution Pixel LArTPC (native 3D) Wire LArTPC (2D projections) Neutrino interaction candidate from MicroBooNE Cosmic rays in a 3D LArTPC charge readout experiment @ Fermilab (arxiv:1808.02969) @ LBNL 4 2018 CPAD / L.Domine and K.Terao

  5. Particle Image Analysis with LArTPCs for neutrinos Neutrino detectors & LArTPCs Goal: Extract flavor + energy 5 2018 CPAD / L.Domine and K.Terao

  6. Particle Image Analysis with LArTPCs for neutrinos Neutrino detectors & LArTPCs Goal: Extract flavor + energy 6 2018 CPAD / L.Domine and K.Terao

  7. Convolutional Neural Networks Now state-of-the art technique in computer vision for complex image analysis tasks: Object detection & classification Semantic segmentation 7 2018 CPAD / L.Domine and K.Terao

  8. Sparse, locally dense data Less than 1% of voxels are nonzero in Sparse (but locally dense) LArTPC images Dense % of nonzero voxels: ~0.05% for 192px^3 ● ~0.01% for 512px^3 ● But CNNs rely on dense matrix multiplications! 8 2018 CPAD / L.Domine and K.Terao

  9. Outline 1. Particle image analysis & Convolutional networks 2. Submanifold Sparse Convolutions 3. Comparison study between a dense and sparse network 9 2018 CPAD / L.Domine and K.Terao

  10. Submanifold Sparse Convolutions Many possible definitions and implementations of ‘sparse convolutions’ ... Submanifold Sparse Convolutions (arxiv:1711.10275, CVPR2018) : https://github.com/facebookresearch/SparseConvNet State-of-the-art on ShapeNet challenge (3D part segmentation) 10 2018 CPAD / L.Domine and K.Terao

  11. Submanifold Sparse Convolutions Submanifold = “input data with lower effective dimension than the space in which it lives” Ex: 1D curve in 2+D space, 2D surface in 3+D space Our case: the worst! 1D curve in 3D space ... 11 2018 CPAD / L.Domine and K.Terao

  12. Submanifold Sparse Convolutions Dilation problem 1 nonzero site leads to 3 d nonzero sites after 1 convolution ● How to keep the same level of sparsity throughout the network? ● 3D Semantic Segmentation with Submanifold Sparse Convolutional Networks (arxiv: 1711.10275) 12 2018 CPAD / L.Domine and K.Terao

  13. Submanifold Sparse Convolutions Predictions Input 2-classes (particle track vs electromagnetic shower ) pixel-level segmentation on 512px 3D images. 13 2018 CPAD / L.Domine and K.Terao

  14. Outline 1. Particle image analysis & Convolutional networks 2. Submanifold Sparse Convolutions 3. Comparison study between a dense and sparse network 14 2018 CPAD / L.Domine and K.Terao

  15. Outline 1. Particle image analysis & Convolutional networks 2. Submanifold Sparse Convolutions 3. Comparison study between a dense and sparse network a. Dataset b. Task c. Metrics d. Network architecture 15 2018 CPAD / L.Domine and K.Terao

  16. 1. Dataset & 2. Task Total: 100,000 simulated 3D events Spatial size: 192px / 512px / 768px (~3mm/pix) Semantic segmentation with 5 classes Protons ● Minimum ionizing particles ● (muons and pions) Electromagnetic shower ● Delta rays ● Michel electrons ● Publicly available: https://osf.io/vruzp/ 16 2018 CPAD / L.Domine and K.Terao

  17. 3. Metrics Nonzero accuracy: fraction of correctly labeled pixels, i.e. ● # nonzero voxels whose predicted label is correct / # nonzero voxels GPU memory (hardware limitation) ● Computation time ● 17 2018 CPAD / L.Domine and K.Terao

  18. 4. Network architecture: UResNet UResNet = U-Net + ResNet (residual connections) Encoder Decoder input conv conv-s2 dconv-s2 linear softmax Residual connections Concatenation 18 2018 CPAD / L.Domine and K.Terao

  19. 4. Network architecture: UResNet UResNet = U-Net + ResNet (residual connections) Encoder Decoder input conv conv-s2 dconv-s2 linear softmax Residual connections Concatenation 19 2018 CPAD / L.Domine and K.Terao

  20. 4. Network architecture: UResNet UResNet = U-Net + ResNet (residual connections) Encoder Decoder input conv conv-s2 dconv-s2 linear softmax Residual connections Concatenation 20 2018 CPAD / L.Domine and K.Terao

  21. 4. Network architecture: UResNet UResNet = U-Net + ResNet (residual connections) Encoder Decoder input Concatenation conv conv-s2 dconv-s2 linear softmax Residual connections Concatenation 21 2018 CPAD / L.Domine and K.Terao

  22. 4. Network architecture: UResNet UResNet = U-Net + ResNet (residual connections) Encoder Decoder input conv conv-s2 dconv-s2 linear softmax Residual connections Residual connections Concatenation 22 2018 CPAD / L.Domine and K.Terao

  23. Semantic Segmentation with UResNet: it works. Data Network’s output A Deep Neural Network for Pixel-Level Electromagnetic Particle Identification in the MicroBooNE Liquid Argon Time Projection Chamber. (arxiv:1808.07269) 23 2018 CPAD / L.Domine and K.Terao

  24. Dense vs Sparse UResNet Dense & Sparse both trained with 80k events Sparse = 99.3% Dense = 92% Nonzero Accuracy (training) vs Iterations 24 2018 CPAD / L.Domine and K.Terao

  25. Dense vs Sparse UResNet Dense & Sparse both trained with 80k events Sparse = 19h Dense = 11 days Nonzero Accuracy (training) vs Wall Time 25 2018 CPAD / L.Domine and K.Terao

  26. Dense vs Sparse UResNet Performance for different input spatial size Sparse Dense Input Spatial Size 192px 512px 768px 192px Final nonzero 98% 98.8% 98.9% 92%* accuracy GPU memory 0.066 0.57 1.0 4.6 usage (Gb) Forward computation 0.058 2.6 3.6 0.68 time (s) *Training time accuracy. 26 2018 CPAD / L.Domine and K.Terao

  27. Dense vs Sparse UResNet Dense = Power? Exponential?! Sparse = Almost linear... 27 2018 CPAD / L.Domine and K.Terao

  28. Learning from mistakes: the case of Michel electrons Nonzero accuracy per class Mean % of Nonzero nonzero voxels accuracy per in an event class = # correctly predicted voxels in this HIP 12% 98.4% class / # voxels in this class MIP 43% 99.5% EM shower 42% 99.1% Delta rays 2% 87.5% Michel 1% 62.8% electrons 28 2018 CPAD / L.Domine and K.Terao

  29. Learning from mistakes: the case of Michel electrons Predictions 29 2018 CPAD / L.Domine and K.Terao

  30. Learning from mistakes: the case of Michel electrons True Labels 30 2018 CPAD / L.Domine and K.Terao

  31. Summary Submanifold sparse convolutions... Run faster ● Use less GPU memory ● … and outperform standard convolutions. ● Better performance and better scalability! Reproduce our results / start using SSCN: Open dataset ● Software containers available ● 31 2018 CPAD / L.Domine and K.Terao

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend