Beyond 2D representa/ons: track/shower separa/on in 3D Ji Won Park - - PowerPoint PPT Presentation

beyond 2d representa ons track shower separa on in 3d
SMART_READER_LITE
LIVE PREVIEW

Beyond 2D representa/ons: track/shower separa/on in 3D Ji Won Park - - PowerPoint PPT Presentation

Beyond 2D representa/ons: track/shower separa/on in 3D Ji Won Park Kazu Terao 11/14/17 SLAC Na/onal Accelerator Laboratory 1 INTRODUCTION 2 Mo/va/ons and goals Long-term mission: build a full 3D reconstruc/on chain for LArTPC data using


slide-1
SLIDE 1

Beyond 2D representa/ons: track/shower separa/on in 3D

Ji Won Park Kazu Terao 11/14/17 SLAC Na/onal Accelerator Laboratory

1

slide-2
SLIDE 2

INTRODUCTION

2

slide-3
SLIDE 3

Mo/va/ons and goals

Goal for the next ~18 minutes: report the results of training a seman/c segmenta/on network to perform track/shower separa/on on 3D simula/on data, as a working test case for paSern recogni/on in 3D.

3

Why in 3D?

  • Less op/cal illusion in interpre/ng 3D data.
  • PID in 2D and track/shower separa/on in 2D have been done

for MicroBooNE data. PaSern recogni/on in 3D is a natural extension from 2D. Long-term mission: build a full 3D reconstruc/on chain for LArTPC data using deep learning.

slide-4
SLIDE 4

Image classifica3on task in 2D Five-par/cle PID has been done: given a 2D image of a single par/cle, label it as a gamma ray, electron, muon, pion, or proton.

4

pion muon From the MicroBooNE CNN paper (2016) gave a happy score distribu3on. *Muon classifica/on score ~ high likely the algorithm thinks an image is a muon. Assigned high scores for muons vs. low scores for pions à confidence in predic/on J

slide-5
SLIDE 5

5

5-par/cle PID in 3D is a natural extension

(achieved similar results as 2D)

pion muon Voxel = the 3D equivalent of pixel

slide-6
SLIDE 6

METHODS

6

slide-7
SLIDE 7

Our study: shower/track separa3on.

(Now 3 classes for track, shower, background instead of 5 par/cle classes)

It has been done pixel-level in 2D using seman/c segmenta/on.

7

Truth label Predic/on From MicroBooNE DNN paper under review Yellow: track, Cyan: shower

slide-8
SLIDE 8

Our study: shower/track separa3on.

(Now 3 classes for track, shower, background instead of 5 par/cle classes)

It has been done pixel-level in 2D using seman/c segmenta/on.

8

Truth label Predic/on From MicroBooNE DNN paper under review Yellow: track, Cyan: shower

Can reuse the network to do shower/track separa/on in 3D. This study allows us to explore how the technique scales to 3D.

slide-9
SLIDE 9

The seman/c segmenta/on network (SSNet)

9

Output Image Input Image Down-sampling Up-sampling

Feature tensor

Intermediate, low-resolu/on feature map

Two components of SSNet

  • 2. Upsampling path
  • 1. Downsampling path

Thank you Kazu for the diagram J

slide-10
SLIDE 10

10

Down-sampling

Feature tensor

  • 1. Downsampling path

Role: classifica3on A series of convolu/ons and downsampling which reduce the input image down to the lowest-resolu/on feature map. Each downsampling step increases field

  • f view of the feature map and allows it

to understand the rela/onship between neighboring pixels. “WriKen texts” input image “Human face” input image “WriKen texts” feature map “Human face” feature map Thank you again Kazu for the diagram J

slide-11
SLIDE 11

11

Up-sampling

Feature tensor

  • 2. Upsampling path

Role: pixel-wise labeling ~ reverse version of downsampling path. A series of convolu/ons-transpose, convolu/ons, and upsampling which retrieve the original resolu/on of the image, with each pixel labeled as

  • ne of the classes.

Segmented output image Each pixel is either “human”

  • r “background”
slide-12
SLIDE 12

12

The type of SSNet we used: U-ResNet

  • 1. Downsampling path
  • 2. Upsampling path

Feature tensor

From MicroBooNE DNN paper under review

slide-13
SLIDE 13

13

The type of SSNet we used: U-ResNet = U-Net + ResNet

From MicroBooNE DNN paper under review

One ResNet module: Within the U-Net architecture, use ResNet modules. In U-ResNet, the convolu/ons are embedded within ResNet modules.

slide-14
SLIDE 14

14

The type of SSNet we used: U-ResNet = U-Net + ResNet

From MicroBooNE DNN paper under review

Concatena/ons: a feature

  • f U-Net.

We stack the feature maps at each downsampling stage with same-size feature maps at the upsampling stage. ~ “shortcut” opera/ons to strengthen correla/on between the low-level details and high-level contextual informa/on.

slide-15
SLIDE 15

Genera/ng images for our training set

15

  • 3D (voxelized)
  • Each event (image) generated

from truth energy deposi/on from LArSoi. With: – Randomized par/cle mul/plicity 1~4 from a unique vertex per event, where the 1~4 par/cles are chosen randomly from 5 par/cle classes. – Momentum varying from 100MeV to 1GeV in isotropic direc/on. – 128 x 128 x 128 voxels à 1cm^3 per voxel (for quick first trial)

Proton 300MeV Proton 360MeV Electron 240MeV Pion 220MeV Input image: each voxel contains charge info.

slide-16
SLIDE 16

Supervised learning: each training example is an ordered pair of input image and true output image (label).

16

Yellow: track, Cyan: shower Label image: each voxel is 0 (background), 1 (shower), or 2 (track).

slide-17
SLIDE 17

Must weight the soWmax cross entropy. Typically, an image has 99.99% background (zero-value) voxels. Even among non-zero voxels, can have uneven number of track voxels vs. shower voxels. So upweight the “rarer” classes in the image, e.g. if the truth label has ra/o of BG: track: shower = 99: 0.7: 0.3, incen/vize the algorithm to do focus

  • n shower most and BG least by

using inverses as weights, 1/99: 1/0.7: 1/0.3.

17

Defining the op/miza/on objec/ve (loss func/on)

Similarly, monitor algorithm’s performance by evalua/ng accuracy only for non-zero pixels

slide-18
SLIDE 18

Training

Op/mizer: Adam Choose batch size to be 8 images

  • batch size ~ size of ensemble, so bigger the beSer BUT

limited by GPU memory

  • one itera/on consumed 8 images

18

slide-19
SLIDE 19

RESULTS

19

slide-20
SLIDE 20

Non-zero pixel accuracy curve

20

Itera/ons Non-zero pixel accuracy = correctly predicted nonzero pixels / total nonzero pixels Each itera/on consumed 8 images. Light orange: raw plot Dark orange: smoothed plot

slide-21
SLIDE 21

Loss curve

21

Itera/ons Each itera/on consumed 8 images. Light orange: raw plot Dark orange: smoothed plot

slide-22
SLIDE 22

Truth label Predic3on

22

slide-23
SLIDE 23

Truth label Predic3on

23

slide-24
SLIDE 24

Truth label Predic3on

24

slide-25
SLIDE 25

Summary and future work

We have trained U-Resnet to perform shower/track separa/on

  • n 3D simula/on data and report a training accuracy of ~96%.

25

To do:

  • Explore smaller voxel sizes for higher precision
  • Vertex finding (adding 1 more class to the classifica/on task)
  • Par/cle clustering (instead of pixel-level, instance-aware

classifica/on)

slide-26
SLIDE 26

BACKUP SLIDES

26

slide-27
SLIDE 27

Overall accuracy curve

27

Itera/ons

slide-28
SLIDE 28

Why ResNet?

This paper demonstrates why ResNet is superior to vgg, etc. in seman/c segmenta/on.

28