Recovering the Full Pose from a Single Keyframe Snowbird, UTAH - - PowerPoint PPT Presentation

recovering the full pose from
SMART_READER_LITE
LIVE PREVIEW

Recovering the Full Pose from a Single Keyframe Snowbird, UTAH - - PowerPoint PPT Presentation

Recovering the Full Pose from a Single Keyframe Snowbird, UTAH 12/2009 Pierre Fite Georgel, Selim Benhimane, Juergen Sotke and Nassir Navab Application Overview Discrepancy check between CAD data and built items ? = 3D Model used as


slide-1
SLIDE 1

Recovering the Full Pose from a Single Keyframe

Snowbird, UTAH 12/2009 Pierre Fite Georgel, Selim Benhimane, Juergen Sotke and Nassir Navab

slide-2
SLIDE 2

Application Overview

  • Discrepancy check between CAD data and built items

3D Model used as Planning for Construction Finished Construction

? = ⇒

2

slide-3
SLIDE 3

Application Overview

  • Support engineer in charge of this verification task
  • Project started in February 2006
  • Consortium of CAMP – Siemens CT – Areva NP

”[…] senior project manager at Siemens, estimates that the software will reduce the cost of constructing a typical medium-sized coal-fired power plant by more than $1m." The economist 2007

3

slide-4
SLIDE 4

Agenda

  • Registration
  • Registration with single keyframe

– Challenges – Initial scale estimates – Non-linear refinement

  • Results

– Synthetic experiments – Plant inspection images

  • Closing statement

4

slide-5
SLIDE 5

Registration Find Geometric Transform between CAD and Image

  • Marker based [1,2]
  • External tracking system

– Magnetic [3] – Optical [4] – GPS + Compass [5]

  • Model based approach

– Edges [6,7] – Keyframes (Multiple [8], Unique with model [9, 10])

[1] Goose et al. Speech-enabled augmented reality supporting mobile industrial maintenance. Pervasive Computing 2003. [2] Pentenrieder et al. Augmented Reality-based factory planning - an application tailored to industrial needs. ISMAR, 2007. [3] Webster et al. Architectural Anatomy. Presence, 1995. [4] Schoenfelder & Schmalstieg. Augmented Reality for Industrial Building Acceptance. IEEE VR, 2008. [5] Schall et al. Virtual redlining for civil engineering in real environments. ISMAR, 2008. [6] Lowe. Fitting parameterized three-dimensional models to images. IEEE Trans. PAMI, 1991. [7] Drummond & Cipolla. Real-time tracking of complex structures with on-line camera calibration. BMVC, 1999. [8] Chia et al. Online 6 dof augmented reality registration from natural features. ISMAR, 2002. [9] Vacchetti et al. Stable real-time 3d tracking using online and offline information. IEEE Trans. PAMI, 2004. [10] Platonov et al. A mobile markerless AR system for maintenance and repair. ISMAR, 2006. 5

slide-6
SLIDE 6

Full Pose from Single Keyframe Challenge

  • Compute fundamental matrix using

keypoints

Target Image Keyframe Local Features Epipolar Lines

F pi, qi q⊤

i Fpi = 0

6

slide-7
SLIDE 7

E

Full Pose from Single Keyframe Challenge

  • Compute fundamental matrix using

keypoints

  • Derivation of essential for calibrated

cameras [11]

Target Image Keyframe Local Features Epipolar Lines

E = K⊤

T FKS

[11] Hung and Faugeras. Some properties of the E matrix in two-view motion estimation. IEEE Trans. PAMI, 1989.

pi, qi q⊤

i Fpi = 0

7

slide-8
SLIDE 8

[11] Hung and Faugeras. Some properties of the E matrix in two-view motion estimation. IEEE Trans. PAMI, 1989.

Full Pose from Single Keyframe Challenge

  • Compute fundamental matrix using

keypoints

  • Derivation of essential for calibrated

cameras [11]

  • Essential matrix decomposition

8 Baseline Direction Target Image Keyframe Local Features Epipolar Lines

E = [t]× R pi, qi

slide-9
SLIDE 9

[11] Hung and Faugeras. Some properties of the E matrix in two-view motion estimation. IEEE Trans. PAMI, 1989.

Full Pose from Single Keyframe Challenge

  • Compute fundamental matrix using

keypoints

  • Derivation of essential for calibrated

cameras [11]

  • Essential matrix decomposition

Unknown translation norm

9 Baseline Direction Target Image Keyframe Local Features Epipolar Lines

E = [t]× R pi, qi

slide-10
SLIDE 10

[11] Hung and Faugeras. Some properties of the E matrix in two-view motion estimation. IEEE Trans. PAMI, 1989.

Full Pose from Single Keyframe Challenge

  • Compute fundamental matrix using

keypoints

  • Derivation of essential for calibrated

cameras [11]

  • Essential matrix decomposition
  • Bundle adjustment cost

10

E = [t]× R CG (Mi, R, t) =

n

  • i=1

+ Ksw (Mi) − pi2 Ktw (RMi + t) − qi2

Baseline Direction Target Image Keyframe

pi qi Mi

pi, qi

slide-11
SLIDE 11

[11] Hung and Faugeras. Some properties of the E matrix in two-view motion estimation. IEEE Trans. PAMI, 1989.

Full Pose from Single Keyframe Challenge

  • Compute fundamental matrix using

keypoints

  • Derivation of essential for calibrated

cameras [11]

  • Essential matrix decomposition
  • Bundle adjustment cost

11

E = [t]× R CG (Mi, R, t) =

n

  • i=1

+ Ksw (Mi) − pi2 Ktw (RMi + t) − qi2 ∀s = 0, CG (sMi, R, st) = CG (Mi, R, t)

Baseline Direction Target Image Keyframe

pi qi Mi

pi, qi

slide-12
SLIDE 12

Full Pose from Single Keyframe Challenge

Baseline Direction Target Image Keyframe

pi qi Mi

12

slide-13
SLIDE 13
  • We suppose that we know and

Full Pose from Single Keyframe Challenge

Baseline Direction Target Image Keyframe

t (t = 1) R

13

slide-14
SLIDE 14
  • We suppose that we know and
  • We search for

Full Pose from Single Keyframe Challenge

Baseline Direction Target Image Keyframe

t (t = 1) R s

14

slide-15
SLIDE 15

Full Pose from Single Keyframe Common Approach

  • Using a known 3D distance D

Baseline Direction Target Image Keyframe

A B d1 = AB 1

15

slide-16
SLIDE 16

Full Pose from Single Keyframe Common Approach

  • Using a known 3D distance D

Baseline Direction Target Image

A B d1 = AB 1 s = D d1

16

slide-17
SLIDE 17

Full Pose from Single Keyframe Common Approach

  • Using the location of a known 3D point in the target image M, q

Baseline Direction Target Image Keyframe

M q

17

slide-18
SLIDE 18

Full Pose from Single Keyframe Common Approach

  • Using the location of a known 3D point in the target image M, q

Baseline Direction Target Image Keyframe

M q s = −

  • K−1

t q

  • × t

⊤ K−1

t q

  • × RM
  • K−1

t q

  • × t
  • 2

18

slide-19
SLIDE 19

Full Pose from Single Keyframe Common Approach

  • Using a known 3D distance
  • Using the location of a known 3D point in the target image

s = −

  • K−1

t q

  • × t

⊤ K−1

t q

  • × RM
  • K−1

t q

  • × t
  • 2

M, q s = D d1 D

Both methods requires interactions

19

slide-20
SLIDE 20
  • We suppose that we know and
  • We search for

Full Pose from Single Keyframe Method Overview

Baseline Direction Target Image Keyframe

t (t = 1) R s

Template Warped Templates Scale Samples

c C l

20

slide-21
SLIDE 21

Full Pose from Single Keyframe Initial Estimates

c C l

  • Every point on a line gives a scale

sample ∀c′ ∈ l, s = −

  • K−1

t c′ × t

⊤ K−1

t c′ × RC

  • K−1

t c′ × t

  • 2

21

slide-22
SLIDE 22

Full Pose from Single Keyframe Initial Estimates

c C l

  • Every point on a line gives a scale

sample

πC n

∀c′ ∈ l, s = −

  • K−1

t c′ × t

⊤ K−1

t c′ × RC

  • K−1

t c′ × t

  • 2

22

slide-23
SLIDE 23

Full Pose from Single Keyframe Initial Estimates

c C l

  • Every point on a line gives a scale

sample

  • We can define a local warping from

the source to the target image

πC n

H (s, πC) = R − stn⊤ d ∀c′ ∈ l, s = −

  • K−1

t c′ × t

⊤ K−1

t c′ × RC

  • K−1

t c′ × t

  • 2

23

slide-24
SLIDE 24

Full Pose from Single Keyframe Initial Estimates

c C l

  • Every point on a line gives a scale

sample

  • We can define a local warping from

the source to the target image

  • Template search

πC n

H (s, πC) = R − stn⊤ d f (s) = SM

  • S, H−1 (s, πC) (T )
  • ∀c′ ∈ l, s = −
  • K−1

t c′ × t

⊤ K−1

t c′ × RC

  • K−1

t c′ × t

  • 2

24

slide-25
SLIDE 25

Full Pose from Single Keyframe Initial Estimates - Overview

… …

Keyframe Template Warped Templates from Target NCC

  • 0.146

0.906 0.437 0.164 0.631 Target Image

25

slide-26
SLIDE 26

[12] Georgel et al. A Unified Approach Combining Photometric and Geometric Information for Pose Estimation. BMVC, 2008.

Full Pose from Single Keyframe Nonlinear Refinement

  • Sub-optimal solution

26

  • i

SM

  • S, H−1 (s, πci) (T )
slide-27
SLIDE 27

[12] Georgel et al. A Unified Approach Combining Photometric and Geometric Information for Pose Estimation. BMVC, 2008.

Full Pose from Single Keyframe Nonlinear Refinement

  • Sub-optimal solution
  • We introduce a quadratic cost

27

CP (R, t) =

m

  • j=1

NCj

  • X

S (Ksw (X)) − T (Ktw (RX + t))2 t ← st [R t]

  • i

SM

  • S, H−1 (s, πci) (T )
  • Ni
slide-28
SLIDE 28

[12] Georgel et al. A Unified Approach Combining Photometric and Geometric Information for Pose Estimation. BMVC, 2008.

Full Pose from Single Keyframe Nonlinear Refinement

  • Sub-optimal solution
  • We introduce a quadratic cost
  • Least square minimization

with

28

CG (Mi, R, t) =

n

  • i=1

+ Ksw (Mi) − pi2 Ktw (RMi + t) − qi2 CP (R, t) =

m

  • j=1

NCj

  • X

S (Ksw (X)) − T (Ktw (RX + t))2 arg min

Mi,R,t CG (Mi, R, t) + CP (R, t)

t ← st [R t]

  • i

SM

  • S, H−1 (s, πci) (T )
  • Ni
slide-29
SLIDE 29

Full Pose from Single Keyframe Algorithm Overview

  • 1. Estimate E
  • 2. Decompose E in R and t
  • 3. for each 2D-3D correspondences

– Find scale s

  • 4. end
  • 5. Select the scale s with best consensus
  • 6. Nonlinear estimation

29

slide-30
SLIDE 30

Full Pose from Single Keyframe Synthetic Experiments

0.01 0.05 0.1 0.5 1 2 0.05 0.1 0.15 0.2 0.25 0.3 2D3D Noise () Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%)

Convergence Rate Reprojection Error before Nonlinear refinement and after 30

slide-31
SLIDE 31

Full Pose from Single Keyframe Synthetic Experiments

0.01 0.05 0.1 0.5 1 2 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 2D2D Noise () Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%)

0.01 0.05 0.1 0.5 1 2 0.05 0.1 0.15 0.2 0.25 0.3 2D3D Noise () Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%)

Convergence Rate Reprojection Error before Nonlinear refinement and after 31

slide-32
SLIDE 32
  • /
+ ++ 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 zoom out / zoom in Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%)

Full Pose from Single Keyframe Synthetic Experiments

3 5 7 15 0.05 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Image Blur (kernel size Px) Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%) 0.01 0.05 0.1 0.5 1 2 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 2D2D Noise () Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%) 0.01 0.05 0.1 0.5 1 2 0.05 0.1 0.15 0.2 0.25 0.3 2D3D Noise () Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%) 1 5 10 25 0.05 0.1 0.15 0.2 0.25 0.3 Image Noise () Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%) 1 2 5 12 24 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 Number of 2D3D Correspondences Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%)

Convergence Rate Reprojection Error before Nonlinear refinement and after 32

slide-33
SLIDE 33
  • /
+ ++ 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 zoom out / zoom in Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%)

Full Pose from Single Keyframe Synthetic Experiments

3 5 7 15 0.05 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 Image Blur (kernel size Px) Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%) 0.01 0.05 0.1 0.5 1 2 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 2D2D Noise () Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%) 0.01 0.05 0.1 0.5 1 2 0.05 0.1 0.15 0.2 0.25 0.3 2D3D Noise () Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%) 1 5 10 25 0.05 0.1 0.15 0.2 0.25 0.3 Image Noise () Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%) 1 2 5 12 24 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1 1.1 1.2 1.3 1.4 1.5 1.6 1.7 Number of 2D3D Correspondences Target Registration Error (Px) 10 20 30 40 50 60 70 80 90 100 Convergence Rate (%)

Stable and precise results Non-linear refinement helps

Convergence Rate Reprojection Error before Nonlinear refinement and after 33

slide-34
SLIDE 34

[13] Georgel et al. An Industrial Augmented Reality Solution For Discrepancy Check. ISMAR, 2007.

Discrepancy Check using Augmented Reality Keyframe Computation

  • Landmark-based registration

– Register images to the CAD coordinate system

  • Approach

– Extract landmarks (Anchor-Plates) from images – Find corresponding 3D landmarks – Compute the pose (Rotation, Translation)

  • Information stored

– Image – Full pose – 3D points: location and normal.

34

slide-35
SLIDE 35

[13] Georgel et al. An Industrial Augmented Reality Solution For Discrepancy Check. ISMAR, 2007.

Discrepancy Check using Augmented Reality Results

35

Robust Features Matches and Propagated 2D-3D Correspondences

slide-36
SLIDE 36

[13] Georgel et al. An Industrial Augmented Reality Solution For Discrepancy Check. ISMAR, 2007.

Discrepancy Check using Augmented Reality Results

36

Resulting Augmentation

slide-37
SLIDE 37

[13] Georgel et al. An Industrial Augmented Reality Solution For Discrepancy Check. ISMAR, 2007.

Discrepancy Check using Augmented Reality Results

37

Robust Features Matches and Propagated 2D-3D Correspondences

slide-38
SLIDE 38

[13] Georgel et al. An Industrial Augmented Reality Solution For Discrepancy Check. ISMAR, 2007.

Discrepancy Check using Augmented Reality Results

38

Resulting Augmentation

slide-39
SLIDE 39

Conclusion and Perspectives  Automatic full pose estimation

 Perspectively corrected template matching  New non-linear cost function

 Solution is used on site for inspection

39

slide-40
SLIDE 40

Conclusion and Perspectives  Automatic full pose estimation

 Perspectively corrected template matching  New non-linear cost function

 Solution is used on site for inspection

  • Extension to more than two images
  • Non calibrated case
  • Estimation of the normals

40

slide-41
SLIDE 41

http://wwwnavab.in.tum.de/Main/PierreGeorgel - Pierre.Georgel@gmail.com

Thank for the Attention

(Any) Questions? 41

slide-42
SLIDE 42

http://wwwnavab.in.tum.de/Main/PierreGeorgel - Pierre.Georgel@gmail.com

42

slide-43
SLIDE 43

Bibliography

[1] Goose et al. Speech-enabled augmented reality supporting mobile industrial maintenance. Pervasive Computing 2003. [2] Pentenrieder et al. Augmented Reality-based factory planning - an application tailored to industrial needs. ISMAR, 2007. [3] Webster et al. Architectural Anatomy. Presence, 1995. [4] Schoenfelder & Schmalstieg. Augmented Reality for Industrial Building Acceptance. IEEE VR, 2008. [5] Schall et al. Virtual redlining for civil engineering in real environments. ISMAR, 2008. [6] Lowe. Fitting parameterized three-dimensional models to images. IEEE Trans. PAMI, 1991. [7] Drummond & Cipolla. Real-time tracking of complex structures with on-line camera calibration. BMVC, 1999. [8] Chia et al. Online 6 dof augmented reality registration from natural features. ISMAR, 2002. [9] Vacchetti et al. Stable real-time 3d tracking using online and offline information. IEEE Trans. PAMI, 2004. [10] Platonov et al. A mobile markerless AR system for maintenance and repair. ISMAR, 2006. [11] Hung and Faugeras. Some properties of the E matrix in two-view motion estimation. IEEE Trans. PAMI, 1989. [12] Georgel et al. A Unified Approach Combining Photometric and Geometric Information for Pose Estimation. BMVC, 2008. [13] Georgel et al. An Industrial Augmented Reality Solution For Discrepancy Check. ISMAR, 2007.

43