3d vision
play

3D Vision Viktor Larsson Spring 2019 Schedule Feb 18 - PowerPoint PPT Presentation

3D Vision Viktor Larsson Spring 2019 Schedule Feb 18 Introduction Feb 25 Geometry, Camera Model, Calibration Mar 4 Features, Tracking / Matching Mar 11 Project Proposals by Students Mar 18 Structure from Motion (SfM) + papers Mar 25


  1. 3D Vision Viktor Larsson Spring 2019

  2. Schedule Feb 18 Introduction Feb 25 Geometry, Camera Model, Calibration Mar 4 Features, Tracking / Matching Mar 11 Project Proposals by Students Mar 18 Structure from Motion (SfM) + papers Mar 25 Dense Correspondence (stereo / optical flow) + papers Apr 1 Bundle Adjustment & SLAM + papers Apr 8 Student Midterm Presentations Apr 15 Multi-View Stereo & Volumetric Modeling + papers Easter break Apr 22 Apr 29 3D Modeling with Depth Sensors + papers May 6 3D Scene Understanding + papers May 13 4D Video & Dynamic Scenes + papers May 20 papers May 27 Student Project Demo Day = Final Presentations

  3. 3D Vision – Class 3 Features & Correspondences feature extraction, image descriptors, feature matching, feature tracking Chapters 4, 8 in Szeliski’s Book [Shi & Tomasi, Good Features to Track, CVPR 1994]

  4. Overview • Local Features • Invariant Feature Detectors • Invariant Descriptors & Matching • Feature Tracking

  5. Importance of Features Features are key component of many 3D Vision algorithms

  6. Importance of Features Schönberger & Frahm, Structure-From-Motion Revisited, CVPR 2016

  7. Feature Detectors & Descriptors • Detector : Find salient structures • Corners, blob-like structures, ... • Keypoints should be repeatable • Descriptor : Compact representation of image region around keypoint • Describes patch around keypoints • Establish matches between images by comparing descriptors

  8. Feature Detectors & Descriptors (Lowe, Distinctive Image Features From Scale-Invariant Keypoints , IJCV’04)

  9. Feature Matching vs. Tracking Matching Tracking • Extract features independently • Extract features in first image • Match by comparing descriptors • Find same feature in next view

  10. Wide Baseline Matching • Requirement to cope with larger variations between images  • Translation, rotation, scaling geometric • Perspective foreshortening transformations  • Non-diffuse reflections photometric transformations • Illumination

  11. Good Detectors & Descriptors? • What are the properties of good detectors and descriptors? • Invariances against transformations • How to design such detectors and descriptors? • This lecture: • Feature detectors & their invariances • Feature descriptors, invariances, & matching • Feature tracking

  12. Overview • Local Features Intro • Invariant Feature Detectors • Invariant Descriptors & Matching • Feature Tracking

  13. Good Feature Detectors? • Desirable properties? • Precise (sub-pixel perfect) localization • Repeatable detections under • Rotation • Translation • Illumination • Perspective distortions • … • Detect distinctive / salient structures

  14. Feature Point Extraction • Find “distinct” keypoints (local image patches) • As different as possible from neighbors homogeneous edge corner

  15. Comparing Image Regions • Compare intensities pixel-by-pixel I ´ (x,y) I(x,y) • Dissimilarity measure: Sum of Squared Differences / Distances ( SSD )     2 SSD  I ( x , y )  I ( x , y )  x y ฀

  16. Finding Stable Features • Measure uniqueness of candidate • Approximate SSD for small displacement Δ    2 SSD  w ( x i ) I ( x i   )  I ( x i ) i     2 w ( x i ) I ( x i )   I  I     I ( x i )      x  y     i   ฀ 2 I x   I I x I x I y   w ( x i )  T      T M   x 2   I x I y I y i ฀ • possible weights ฀ ฀

  17. Finding Stable Features homogeneous edge corner Suitable feature positions should maximize i.e. maximize smallest eigenvalue of M

  18. Harris Corner Detector • Use small local window: • Directly computing eigenvalues λ 1 , λ 2 of M is computationally expensive • Alternative measure for “ cornerness ”: = 𝜇 1 ⋅ 𝜇 2 − 𝑙 𝜇 1 + 𝜇 2 2 • Homogeneous: 𝜇 1 , 𝜇 2 small ⇒ 𝑆 small 2 < 0 • Edge: 𝜇 1 ≫ 𝜇 2 ≈ 0 ⇒ 𝑆 = 𝜇 1 ⋅ 0 − 𝑙𝜇 1 • Corner: 𝜇 1 , 𝜇 2 large ⇒ 𝑆 large

  19. Harris Corner Detector • Alternative measure for “ cornerness ” • Select local maxima as keypoints • Subpixel accuracy through second order surface fitting (parabola in 1D)

  20. Harris Corner Detector • Keypoint detection: Select strongest features over whole image or over each tile (e.g. 1000 per image or 2 per tile) • Invariances against geometric transformations • Shift / translation?

  21. Geometric Invariances Rotation Harris: Yes Scale Harris: No Affine (approximately invariant w.r.t. perspective/viewpoint) Harris: No

  22. 2D Transformations of a Patch Harris corners VIP Harris corners MSER SIFT

  23. Scale-Invariant Feature Transform (SIFT) • Detector + descriptor (later) • Recover features with position, orientation and scale (Lowe, Distinctive Image Features From Scale-Invariant Keypoints , IJCV’04)

  24. Position • Look for strong responses of Difference-of- Gaussian filter ( DoG ) 3 2  k • Approximates Laplacian of Gaussian ( LoG ) • Detects blob-like structures • Only consider local extrema

  25. Scale • Look for strong DoG responses over scale space    ฀  1/2 image ( σ =2) ฀ ฀ ฀    ฀  ฀ orig. image   4 ฀ 2 Slide credits: Bastian Leibe, Krystian Mikolajczyk ฀ ฀

  26. Scale • Only consider local maxima/minima in both position and scale • Fit quadratic around extrema for sub- pixel & sub-scale accuracy

  27. Minimum Contrast and “ Cornerness ” all features

  28. Minimum Contrast and “ Cornerness ” after suppressing edge-like features

  29. Minimum Contrast and “ Cornerness ” after suppressing edge-like features + small contrast features

  30. Invariants So Far • Translation? Yes • Scale? Yes • Rotation? Yes

  31. Orientation Assignment • Compute gradient for each pixel in patch at selected scale • Bin gradients in histogram & smooth histogram • Select canonical orientation at peak(s) • Keypoint = 4D coordinate 2  0 (x, y, scale, orientation)

  32. Invariants So Far • Translation • Scale • Rotation • Brightness changes: • Additive changes? • Multiplicative changes?

  33. 2D Transformations of a Patch Harris corners VIP Harris corners MSER SIFT

  34. Affine Invariant Features Perspective effects can locally be approximated by affine transformation

  35. Extreme Wide Baseline Matching • Detect stable keypoints using the Maximally Stable Extremal Regions ( MSER ) detector • Detections are regions , not points! (Matas et al., Robust Wide Baseline Stereo from Maximally Stable Extremal Regions, BMVC’02)

  36. Maximally Stable Extremal Regions Extremal regions: • Much brighter than surrounding • Use intensity threshold

  37. Maximally Stable Extremal Regions Extremal regions: • OR: Much darker than surrounding • Use intensity threshold

  38. Maximally Stable Extremal Regions • Regions: Connected components at a threshold • Region size = #pixels • Maximally stable: Region constant near some threshold

  39. A Sample Feature

  40. A Sample Feature T is maximally stable wrt. surrounding

  41. From Regions To Ellipses • Compute „ center of gravity “ • Compute Scatter (PCA / Ellipsoid)

  42. From Regions To Ellipses • Ellipse abstracts from pixels! • Geometric representation: position/size/shape

  43. Achieving Invariance • Normalize to „ default “ position, size, shape • For example: Circle of radius 16 pixels

  44. • Normalize ellipse to circle (affine transformation) • 2D rotation still unresolved

  45. • Same approach as for SIFT: Compute histogram of local gradients • Find dominant orientation in histogram • Rotate local patch into dominant orientation

  46. Summary: MSER Features • Detect sets of pixels brighter/darker than surrounding pixels • Fit elliptical shape to pixel set • Warp image so that ellipse becomes circle • Rotate to dominant gradient direction (other constructions possible as well)

  47. MSER Features - Invariants • Constant brightness changes (additive and multiplicative) • Rotation, translation, scale • Affine transformations  Affine normalization of feature leads to similar patches in different views !

  48. 2D Transformations of a Patch Harris corners VIP In practice hardly observable for small patches ! Harris corners MSER SIFT

  49. Viewpoint Invariant Patches (VIP) • Use known planar geometry to remove perspective distortion • Or: Use vanishing points to rectify patch (Wu et al., 3D Model Matching with Viewpoint Invariant Patches (VIPs), CVPR’08)

  50. Learning Feature Detectors • In the age of deep learning, can we learn good detectors from data? • How can we model repeatable feature detection? • Learn ranking function H(x|w): R 2 → [ -1, 1] with parameters w • Interesting points close to -1 or 1 (Savinov et al., Quad-networks: unsupervised learning to rank for interest point detection, CVPR’17)

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend