Light Field Vision for Transparent Object Categorization and - - PowerPoint PPT Presentation

light field vision for transparent object categorization
SMART_READER_LITE
LIVE PREVIEW

Light Field Vision for Transparent Object Categorization and - - PowerPoint PPT Presentation

Light Field Vision for Transparent Object Categorization and Segmentation Yichao Xu xuyichao.cn Jan. 6, 2016 Just a reminder Last day P4A-04 1 About me A: Hometown


slide-1
SLIDE 1

Light Field Vision for Transparent Object Categorization and Segmentation 光场视觉在透明物体分类和分割中的应用

Yichao Xu 徐轶超

  • Jan. 6, 2016

xuyichao.cn

slide-2
SLIDE 2

Just a reminder – Last day P4A-04

1

slide-3
SLIDE 3

About me

  • A: Hometown in Zhejiang
  • Jiaxing
  • B: Undergraduate in Beijing
  • BESTI
  • C: Master 1 in Anhui
  • USTC, Hefei
  • D: Master 2-3 in Shanghai
  • SINAP

, CAS

  • E: PhD in Fukuoka, Japan
  • Kyushu University

2

slide-4
SLIDE 4

Outline

  • Introduction of Light Field Vision
  • Transcat: Transparent Object Categorization
  • Transcut: Transparent Object Segmentation

3

slide-5
SLIDE 5

Scene

Light field

4 Light field describes all the light rays in the space

slide-6
SLIDE 6

Sensors for visual perception

5 Cameras with CCD and CMOS sensors

slide-7
SLIDE 7

Only a few light rays can be captured

Scene Image

Regular camera sensing

6

slide-8
SLIDE 8

Each light ray can be represented by L(s, t, u, v)

Scene

Light field parameterization

7

Position (s, t) Angle (u, v)

4D light field

slide-9
SLIDE 9

u 𝑡 Light field camera can capture richer information

Scene Viewpoint plane Sensor plane

Light field sensing

8

slide-10
SLIDE 10

u 𝑡 Regular camera can only sample sub light field space

Scene Viewpoint plane Sensor plane

Light field sampling in phase space

9 u 𝑡

𝑡u phase space

slide-11
SLIDE 11

u 𝑡 Light field camera can capture richer information

Scene Viewpoint plane Sensor plane

Light field sampling in phase space

10 u 𝑡

𝑡u phase space

slide-12
SLIDE 12

Computational Photography

11 Multi-focus Multi-view Light Field is widely used for Image-based Rendering

slide-13
SLIDE 13

12

Simultaneously record positional and angular information of ray

Obtain rich information with single-shot

Light field cameras

Stanford Raytrix Viewplus Lytro PiCam

slide-14
SLIDE 14

Light field vision

13 Capture To solve computer vision problems

slide-15
SLIDE 15

Computer vision makes our life better

14 Help us know more Free our hands

slide-16
SLIDE 16

Visual recognition makes it possible

15 Visual recognition is important in these applications France Prešeren, Poet

slide-17
SLIDE 17

Advantage of light field vision

16 Regular Computer vision Light Field Vision

Redundant information makes it easier to understand the 3D world

slide-18
SLIDE 18

Light field vision applications

  • Surveillance - Accurately detect desired foreground
  • Depth estimation - Accurate and consistent
  • Salience detection - Accurate in challenge scenes

17

[A.Shimada et al., IPSJ CVA 2013] LF method Conventional LF method Conventional [S. Wanner et al., PAMI2014] [N. Li et al., CVPR2014] LF method Conventional GT

slide-19
SLIDE 19

Light Field Vision Application

  • - transparent object recognition

18

slide-20
SLIDE 20

Transparent object recognition

19 Which type is the object? Where is the object?

slide-21
SLIDE 21

Challenge of the target object

20

Appearance of transparent objects drastically varies with background

slide-22
SLIDE 22

Transparent object causes distortion

Regular computer vision methods cannot understand whether the image is distorted or not without prior knowledge Different objects produce different image of the same scene 21

slide-23
SLIDE 23

Know light field from background

22 [Ben-Ezra and Nayar, ICCV2003] Known motion, Manually tagged feature [G. Wetzstein et al, ICCV2011] Known background Transparent object

slide-24
SLIDE 24

Features from Light Field for Transparent Object Recognition

23

slide-25
SLIDE 25

Distortion modeled by light field vision

Background distortion changes with viewpoint Background distortion is modeled as the correspondences between the viewpoints 24

slide-26
SLIDE 26

Modeled distortion is independent of background textures

Background invariant distortion

25

slide-27
SLIDE 27

∆u ∆v

Light Field Distortion (LFD) feature

26

slide-28
SLIDE 28

LFD feature visualization

27 2D vectors on different viewpoints 24x2D feature vector for each pixel

∆u ∆v

slide-29
SLIDE 29

u u 𝑡 𝑡 Rays from background are linear distributed

Background Viewpoint plane Sensor plane 𝑡u phase space

Light Field Linearity (LF-linearity)

28

slide-30
SLIDE 30

Transparent object

𝑡

Sensor plane Background Viewpoint plane

u u 𝑡

𝑡u phase space

Light Field Linearity (LF-linearity)

Rays from transparent object are not linear distributed 29

slide-31
SLIDE 31

Extract LF-linearity

30

∆𝑣 𝑡 Disparity Euclidean Distance Hyper-plane ∆𝑣 𝑡

slide-32
SLIDE 32

LF-linearity visualization

31 Central view LF-linearity

slide-33
SLIDE 33

Light Field Consistency (LF-consistency)

( , ) view s t (0,0) view

forward matching backward matching

Good consistency Poor consistency 32 LF-consistency is used for detecting the depth discontinuity

forward matching backward matching

(0,0) view ( , ) view s t

slide-34
SLIDE 34

Occlusion in light field

Occlusion is caused by depth discontinuity 33 Occlusion detector

slide-35
SLIDE 35

Occlusion detectors

34

slide-36
SLIDE 36

Detect occlusion point

35 X = 1 The detected occlusion point is from θ = 0

1 1 1 1 1 1 1 1 1 1 1

= 0.7

slide-37
SLIDE 37

Detected occlusion visualization

36 Central view Occlusion response

slide-38
SLIDE 38

Feature and descriptor

  • LFD Feature (光场扭曲特征)
  • 2x24 Dimensional vector
  • Describe the distortion pattern
  • LF-linearity(光场线性度)
  • A metric to describe how much is the distortion
  • Occlusion detector (遮挡检测)
  • Describe the probability of

a point to be in the occlusion

  • Occlusion in which direction

37

slide-39
SLIDE 39

Outline

  • Introduction of Light Field Vision
  • Transcat: Transparent Object Categorization
  • Transcut: Transparent Object Segmentation

38

slide-40
SLIDE 40

TransCat: Transparent Object Categorization

39 Which object?

slide-41
SLIDE 41

Training pipeline

Extracting the LFD feature Training based on Bag of features Training based on Bag of features Estimate background by LF-linearity

40

Filtering the background Extracting the LFD feature

Non-linear Non-linear Non-linear Non-linear Linear Linear Linear Linear

slide-42
SLIDE 42

Extracting the LFD feature Filtering the background Training based on Bag of features

Training pipeline

41

Estimate relative differences by

  • ptical flow
slide-43
SLIDE 43

Representative LFD Feature space Extracting the LFD feature Filtering the background by LF-linearity Training based on Bag of features

Training pipeline

42

slide-44
SLIDE 44

Representative LFD

Categorization based

  • n Bag of features

Testing for transparent objects categorization

43

result

slide-45
SLIDE 45

Experimental setting

18 objects 10 backgrounds 44

Background scenes can be dynamic!

slide-46
SLIDE 46

Categorization result

0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 A B C D E F G H I J K L M N O P Q R

Recognition ratio Object

Average categorization accuracy: 84% 45

Evaluation by leave-one-out cross-validation

slide-47
SLIDE 47

Analysis

  • Applicable conditions

46

0.2 0.4 0.6 0.8 1 30 35 40 45 50 Recognition ratio Camera position [cm] 0.2 0.4 0.6 0.8 1 50 100 150 200 250 Recognition ratio Background position [cm] 0.2 0.4 0.6 0.8 Recognition ratio Lighting angle [degree] 0.2 0.4 0.6 0.8 0.1 0.2 0.3 Recognition ratio Noise standard deviation

slide-48
SLIDE 48

Analysis

  • Applicable conditions

47

0.2 0.4 0.6 0.8 1

  • 10
  • 5

5 10 Recognition ratio Rotation along x-axis [degree] 0.2 0.4 0.6 0.8 10 20 30 40 Recognition ratio Rotation along y-axis [degree] 0.2 0.4 0.6 0.8 10 20 30 40 Recognition ratio Rotation along z-axis [degree]

Overall Symmetric objects Asymmetric objects

slide-49
SLIDE 49

Results for real scene

48

slide-50
SLIDE 50

Outline

  • Introduction of Light Field Vision
  • Transcat: Transparent Object Categorization
  • Transcut: Transparent Object Segmentation

49

slide-51
SLIDE 51

Transcut: Transparent Object Segmentation

50

slide-52
SLIDE 52

Properties of different components

Transparent object segmentation formulated as labeling problem 51

Trans Obj Background Good LF-linearity Occlusion Extracted by occlusion detector Transparent Object Poor LF-linearity exclude the occlusion

slide-53
SLIDE 53

Regional term

Central view of input light field image

Background penalty Foreground penalty

large penalty assigns to pixels that have poor LF-linearity exclude the occlusion area large penalty assigns to pixels with poor LF-linearity in the occlusion or pixels with good LF-linearity 52

slide-54
SLIDE 54

Boundary term

Central view of input light field image

q3

… …

p q2 q4 q1

… …

If is from 𝑃𝑞 θ = 0,

Detected

  • cclusion point

𝑃𝑞 53

slide-55
SLIDE 55

Energy minimization

54 Central view of input light field image Regional term Boundary term

Graph Cut

slide-56
SLIDE 56

Experiments

Object 1 Object 2 Object 3 Object 4 Object 5 Object 6 Object 7 Scene 1 Scene 2 Scene 3 Scene 4 Scene 5 Scene 6 Scene 7

55 Background scenes can be dynamic!

slide-57
SLIDE 57

Comparison with related work

56

Finding glass

McHenry et al., CVPR2005

6 features from appearance LF-linearity thresholding Single feature from LF

slide-58
SLIDE 58

Visual comparison

57

Images from the central viewpoint Results from Finding glass Results from LF-linearity thresholding Results from TransCut Ground Truth

slide-59
SLIDE 59

Quantitative comparison

58

TP FN FP

F-measure Recall Precision Th=1 0.19 0.84 0.11 Th=2 0.34 0.78 0.22 Th=3 0.42 0.76 0.29 Th=4 0.45 0.73 0.33 Th=5 0.48 0.70 0.37 Th=6 0.49 0.67 0.39 Th=7 0.50 0.65 0.41 Th=8 0.50 0.63 0.43 Th=9 0.50 0.61 0.44 F-measure Recall Precision Finding glass 0.30 0.82 0.19 LF-linearity thresholding 0.50 0.65 0.41 Proposed 0.85 0.96 0.77

slide-60
SLIDE 60

Summary

  • Light field vision can get more

information for solving vision problems

  • Full space sampling vs. sub-space

sampling

  • Transparent object categorization
  • Transparent object segmentation

59

slide-61
SLIDE 61

Open issues

  • Develop robust feature descriptors
  • Distance invariant
  • Rotation invariant
  • Apply to other objects
  • Specular objects
  • Recover the undistorted background
  • Reconstruct 3D shape of transparent objects
  • Natural scenes

60

slide-62
SLIDE 62

Publications

  • Y. Xu, K. Maeno, H. Nagahara, and R. Taniguchi, “Mobile camera

array calibration for light field acquisition,” in International Conference on Quality Control by Artificial Vision (QCAV), pp. 283–290, 2013.

  • Y. Xu, K. Maeno, H. Nagahara, and R. Taniguchi, “Camera array

calibration for light field acquisition,” Frontiers of Computer Science, 2015, 9(5), pp. 691-702.

  • Y. Xu, K. Maeno, H. Nagahara, A. Shimada, and R. Taniguchi,

“Light field distortion feature for transparent object classification,” Computer Vision and Image Understanding, Vol. 139, pp. 122-135, 2015.

  • Y. Xu, H. Nagahara, A. Shimada, and R. Taniguchi, "TransCut:

Transparent Object Segmentation from a Light-Field Image", ICCV 2015, Santiago, Chile

61

slide-63
SLIDE 63

Thank you!

62