Bounding Box Regression With Uncertainty for Accurate Object - - PowerPoint PPT Presentation

bounding box regression with uncertainty for accurate
SMART_READER_LITE
LIVE PREVIEW

Bounding Box Regression With Uncertainty for Accurate Object - - PowerPoint PPT Presentation

Bounding Box Regression With Uncertainty for Accurate Object Detection 1 Carnegie Mellon University 2 Megvii Yihui He 1 , Chenchen Zhu 1 , Jianren Wang 1 , Marios Savvides, 2 Xiangyu Zhang Ambiguity: inaccurate labelling MS-COCO Ambiguity:


slide-1
SLIDE 1

Bounding Box Regression With Uncertainty for Accurate Object Detection

1Carnegie Mellon University 2Megvii

Yihui He1, Chenchen Zhu1, Jianren Wang1, Marios Savvides, 2Xiangyu Zhang

slide-2
SLIDE 2

Ambiguity: inaccurate labelling

  • MS-COCO
slide-3
SLIDE 3

Ambiguity: inaccurate labelling

  • MS-COCO
slide-4
SLIDE 4

Ambiguity: introduced by occlusion

  • MS-COCO
slide-5
SLIDE 5

Ambiguity: object boundary itself is ambiguous

  • YouTube-BoundingBoxes
slide-6
SLIDE 6

Classification Score & Localization misalignment

MS-COCO VGG-16 Faster RCNN

slide-7
SLIDE 7

Standard Faster R-CNN Pipeline

1024 x 81 1024 x 81x4

Cross entropy/focal loss

slide-8
SLIDE 8

Modeling bounding box prediction

  • Predict Gaussian distribution instead of a number

https://upload.wikimedia.org/wikipedia/commons/9/9e/Normal_Distribution_NIST.gif

slide-9
SLIDE 9

Modeling ground truth bounding box

  • Dirac delta function

https://upload.wikimedia.org/wikipedia/commons/b/b4/Dirac_function_approximation.gif

slide-10
SLIDE 10

KL Loss: Gaussian meets delta function

slide-11
SLIDE 11

Architecture

An additional fully-connected layer for prediction variance (1024 x 81 x 4) 1024 x 81 1024 x 81x4 1024 x 81x4

slide-12
SLIDE 12

Why KL Loss

(1) The ambiguities in a dataset can be successfully captured. The bounding box regressor gets smaller loss from ambiguous bounding boxes. (2) The learned variance is useful during post-processing. We propose var voting (variance voting) to vote the location of a candidate box using its neighbors’ locations weighted by the predicted variances during nonmaximum suppression (NMS). (3) The learned probability distribution is interpretable. Since it reflects the level of uncertainty of the bounding box prediction, it can potentially be helpful in down-stream applications like self-driving cars and robotics

slide-13
SLIDE 13

KL Loss: Degradation Case

slide-14
SLIDE 14

KL Loss: Reparameterization trick

convert α back to σ during testing

slide-15
SLIDE 15

KL Loss: Rubust L1 Loss (Smooth L1 Loss)

Smooth L1 Loss KL Loss

slide-16
SLIDE 16

KL Loss: Uncertainty Prediction

Sigma in Green box

slide-17
SLIDE 17

KL Loss: Uncertainty Prediction

Sigma in Green box

slide-18
SLIDE 18

KL Loss: Uncertainty Prediction

Sigma in Green box

slide-19
SLIDE 19

KL Loss: Uncertainty Prediction

Sigma in Green box

slide-20
SLIDE 20

Variance Voting

  • Larger IoU gets higher score
  • Lower variance gets higher score
  • Classification score invariance
slide-21
SLIDE 21

Variance Voting

Before after

slide-22
SLIDE 22

Variance Voting

Before after

slide-23
SLIDE 23

Variance Voting

Before after

slide-24
SLIDE 24

Variance Voting

Before after

slide-25
SLIDE 25

Ablation Study: KL Loss, soft-NMS, Variance Voting

  • VGG-16
  • MS-COCO
slide-26
SLIDE 26

Ablation Study: does #params in head matter?

The Larger R-CNN head, the better

slide-27
SLIDE 27

Ablation Study: Variance Voting Threshold

σt = 0, standard NMS Large σt: farther boxes are considered

slide-28
SLIDE 28

Improving State-of-the-Art

  • Mask R-CNN
  • MS-COCO
slide-29
SLIDE 29

Inference Latency

  • VGG-16
  • single image
  • single GTX 1080 Ti GPU

2ms

slide-30
SLIDE 30

Other models on MS-COCO

slide-31
SLIDE 31

VGG on PASCAL VOC

slide-32
SLIDE 32

Join us at Tuesday Afternoon Poster Session #41

Bounding Box Regression with Uncertainty for Accurate Object Detection