cnn ba cnn based ed pi pipeline peline for or
play

CNN Ba CNN Based ed Pi Pipeline peline for or Op Optical ical - PowerPoint PPT Presentation

CNN Ba CNN Based ed Pi Pipeline peline for or Op Optical ical Fl Flow ow Tal Schuster, June 2017 Based on: PatchBatch: a Batch Augmented Loss for Optical Flow, (Gadot, Wolf) CVPR 2016 Optical Flow Requires Multiple Strategies (but only


  1. CNN Ba CNN Based ed Pi Pipeline peline for or Op Optical ical Fl Flow ow Tal Schuster, June 2017 Based on: PatchBatch: a Batch Augmented Loss for Optical Flow, (Gadot, Wolf) CVPR 2016 Optical Flow Requires Multiple Strategies (but only one network), (Schuster, Wolf, Gadot) CVPR 2017 1

  2. Overv rview iew Goal – Get SOTA results in main optical flow benchmarks Was done by: ● Constructing a Deep Learning based pipeline (modular) ● Architectures exploration ● Loss function augmentations ● Per-batch statistics ● Learning methods 2

  3. Problem Definition 3

  4. Problem blem Defi finiti nition on - Optic ical al Flow ow Given 2 images, compute a dense Optical Flow Field describing the motion between both images (i.e. pure re optical flow): 2 X (h,w,1 / 3) → (h,w,2) Where: ● h - image height, w - image width ● (h,w,1 / 3) - a grayscale or RGB image ● (h,w,2) - a 3D tensor describing for each point (x,y) in image-A a 2D-flow vector: ( Δ𝑦, Δ𝑧 ) Accuracy measures: ● Based on GT (synthetic or physically obtained) - KITTI, MPI-Sintel ● F_err - % of pixels with euclidean error > z pixels (usually z=3) ● Avg_err - mean of euclidean errors over all pixels 4

  5. DB DB - KITTI TTI20 2012 12 ● LIDAR based ● ~50% coverage 5

  6. DB DB - KITTI TTI20 2012 12 6

  7. DB DB - KITTI TTI20 2015 15 7

  8. DB DB - MPI SINTEL NTEL ● Synthetic (computer graphics) ● ~100% coverage 8

  9. Solutions tions Traditional computer vision methods ● Global constraints (Horn-Schunk, 1981) – Brightness constancy + smoothness asm. ● Local constraints (Lucas-Kanade, 1981) Main disadvantage – small objects and fast movements Descriptor based methods ● Sparse to dense (Brox-Malik, 2010) Descriptors SIFT, SURF, HOG, DAISY, etc. (handcrafted) CNN methods ● End to End – Flownet (Fischer et al., 2015) 9

  10. Refer erenc ence e Work k – Zbontar ontar & & Lecun, n, 2015 2015 Solving Stereo-Matching vs. Optical Flow Classification-based vs. metric learning To compute the classification, the network needs to observe both patches simultaneously 10

  11. The PatchBatch pipeline 11

  12. PatchBatch hBatch - DNN DNN Siamese DNN - i.e., tied weights due to symmetry Leaky ReLU Should be FAST: Matching function = L2 Conv only Independent descriptor computation 12

  13. PatchBatch hBatch - Overall all Pipel eline ine (Normalized) Keeping only large connected components PatchMatch - Barnes et al. 2010 EpicFlow - Revaud et at. 2015 13

  14. PatchBatch hBatch - ANN ANN PatchMatch: (Descriptors, Matching function) → ANN ANN and not ENN : O(N^2) → O(N*logN) 2 iterations are enough 1. Initialization (random) 2. Propagation 𝑔 𝑦, 𝑧 = 𝑏𝑠𝑕𝑛𝑗𝑜 𝐸 𝑔 𝑦, 𝑧 , 𝐸 𝑔 𝑦 − 1, 𝑧 , 𝐸 𝑔 𝑦, 𝑧 − 1 (+1 on even iterations) 3. Search 𝑆 𝑗 ∈ −1,1 × [−1,1] 𝑣 𝑗 = 𝑤 0 + 𝑥𝛽 𝑗 𝑆 𝑗 𝑥 - max radius 4. Return to step 2 1 𝛽 - step ( = 2 ) 14

  15. PatchBatch hBatch - Post Post-Proces Processi sing ng EpicFlow (Edge-Preserving Interpolation of Correspondences) Sparse -> Dense Average support affine transformations based on geodesic distance on top of edges map SED alg. 15

  16. PatchBatch hBatch - CNN NN Batch Normalization- Solves the “ internal covariate shift ” problem Per pixel instead of per feature map 16

  17. PatchBatch hBatch - Loss Loss ● DrLIM - Dimensionality Reduction by Learning an Invariant Mapping (LeCun, 2006) Orig DrLIM (SPRING) CENT CENT+SD Negative pairs 𝑀 𝐸 𝓍 CENT CENT + SD 17

  18. PatchBatch hBatch - Trai aining ning Method hod Negative sample – random 1-8 pixels from the true match Data augmentation - flipping , rotating 90 ° 18

  19. Results 19

  20. Benchmar hmarks 20

  21. How can we Improv prove e the Results lts? 21

  22. Architecture Modifications 22

  23. PatchBatch hBatch - CNN NN Increased Patch and Descriptor sizes 23

  24. Hinge ge Loss s with h SD • Hinge Loss s instead of DrLIM • Trained on Triplets ts - <A, B-match, B-non-match > • Keeping the additional SD c comp mponent 24

  25. Failed led Attempts empts Data augmentation ● Rotations (random +/- 𝛽 ) ● Colored patches (HSV or other decomposition) Loss function ● Foursome (A, B, A ’ , B ’ ) A, B – matching patches A ’– Patch from Image A that is closest to B H(A,B) = max(0, m - 𝑀 2 (A,B)) L = 𝑀 2 (A,B) + Ι( B ≠ B ’ ) * H(A,B ’ ) + Ι( A ≠ A ’ ) * H(A ’ ,B) Sample Mining ● PatchMatch output ● Patches distance ● Descriptor distance 25

  26. Optical Flow as a Multifaceted Problem 26

  27. Success ess of methods hods MPI-Sintel results table The challenge of larg rge disp splace ceme ments ts KITTI 2015 average error: ● Foreground – 26.43 % ● Background – 11.43 % Possible causes: ● Matching algorithm ● Descriptors quality PatchBatch on KITTI 2012 – distance between true matches 27

  28. Descri criptor ptors Evaluation uation Defined a quality measurement of descriptors for matching 𝑒 𝑞 is a distra tractor tor of pixel 𝑞 if the 𝑀 2 distance between 𝑒 𝑞 and 𝑞 is lower than the 𝑀 2 distance of 𝑞 with its matching pixel descriptor. Counted up to 25 pixels from the examined pixel. 28

  29. Distrac tractor tors by displacement lacement Amount of distractors increase with displacement range Goal: improve results for large displacements without reducing for other ranges. 29

  30. Distrac tractor tors by displacement lacement Amount of distractors increase with displacement range Goal: improve results for large displacements without reducing for other ranges. Expert models ● Training only on sub ranges ● Improving results for large displacements is possible ● Implies the need of differ ferent t feature tures for differe ferent patch ches 30

  31. Learning with Varying Difficulty 31

  32. Gradual dual Learni arning ng Methods hods Deal with varying difficulty Curriculum (Bengio et al. 2009): ● Samples are pre-ordered ● Curriculum by displacement ● Curriculum by distance (of false sample) Self-Paced (Kumar et al. 2010): ● No need to pre-order ● Sample hardness increases with time (by loss value) Hard Mining (Simo-Serra et al. 2015): ● Backpropagate only some ratio of harder samples ● Used for training local descriptors with triplets All methods did not improve match over baseline – Why? 32

  33. Need ed for r variant riant extractin tracting g strategies rategies Large motions are mostly correlated with more changes in appearance: 1. Background changes 2. View point changes -> occluded parts 3. Distance and angle to light source -> illumination 4. Scale (when moving along the Z-axis) 34

  34. Learning for Multiple Strategies and Varying Difficulty 35

  35. Our Interleav erleaving ing Learni arning ng Method hod Goal: Deal with mult ltiple iple sub-task tasks Classification: Painting to Artist Massing Interleaving Learning ML models ● Mostly in random order (SGD) ● Applying gradual methods can effect randomness Interl rleavi ving Learn rning ● Maintai taining the random m order of categories s while adjust sting the diffic ficulty ty Motivated by psychological research (Kornell - Bjork) ● Massing vs. Interleaving ● Experiments on classification tasks, sports, etc. Learning Concepts and Categories – Kornell and Bjork (2008) 36

  36. Same e class s of objec ects? ts? 37

  37. Interleavi erleaving ng Learning rning for Optical al Flow Controlling the negative sample to balance difficulty Original method Interleaving 38

  38. Interleavi erleaving ng Learning rning for Optical al Flow ● Drawing the line from 𝑞 improved matching results but did not effect the distractors measurement (Due to PatchMatch initialization) 39

  39. Self Self-Pac Paced ed Curric rriculum ulum Interleav erleaving ing Learning rning (SPCI) CI) 𝑚 𝑗 - validation loss on epoch 𝑗 𝑚 𝑗𝑜𝑗𝑢 - initial loss value (epoch #5) 𝑛 – total epoch amount 40

  40. Experiments 41

  41. Optical cal Flow 42

  42. Results 43

  43. Benchmar hmarks s - KITTI TTI20 2012 12 44

  44. Benchmar hmarks s - KITTI TTI20 2015 15 45

  45. 46

  46. 47

  47. 48

  48. Benchmar hmarks s - MPI MPI-Sintel Sintel 49

  49. 50

  50. 51

  51. Summary 52

  52. Summary mary ● Computing Optical Flow as a matching problem with a modular pipeline ● using a CNN to generate descriptors ● Per-batch statistics (SD, batch normalization) ● Interleaving Learning Method & SPCI Referring difficulties while maintaining a random order of the categories ● One model to generate descriptors for both small and large displacements 53

  53. THANK YOU! 54

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend