Visual Perception Sensors Depth Determination Gerrit Glaser - - PowerPoint PPT Presentation

visual perception sensors
SMART_READER_LITE
LIVE PREVIEW

Visual Perception Sensors Depth Determination Gerrit Glaser - - PowerPoint PPT Presentation

MIN Faculty Department of Informatics Visual Perception Sensors Depth Determination Gerrit Glaser University of Hamburg Faculty of Mathematics, Informatics and Natural Sciences Department of Informatics Technical Aspects of Multimodal


slide-1
SLIDE 1

MIN Faculty Department of Informatics

Visual Perception Sensors

Depth Determination Gerrit Glaser

University of Hamburg Faculty of Mathematics, Informatics and Natural Sciences Department of Informatics Technical Aspects of Multimodal Systems

November 13. 2017

  • G. Glaser – Visual Perception Sensors

1 / 27

slide-2
SLIDE 2

Table of Contents

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

  • 1. Motivation

Camera

  • 2. Triangulation Approaches

Stereoscopic Cameras Binary Projection Camera Microsoft Kinect

  • 3. Time of Flight Approaches

Depth Camera Kinect V2 LIDAR

  • 4. Conclusion
  • G. Glaser – Visual Perception Sensors

2 / 27

slide-3
SLIDE 3

Motivation for Visual Perception in Robotics

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ basic question for mobile robotics: Where am I? ◮ autonomous movement through unknown terrain

◮ scan environment for obstacles ◮ distances to surroundings

Possible solution

Add visual perception sensors, to allow robots to “see” their environment.

  • G. Glaser – Visual Perception Sensors

3 / 27

slide-4
SLIDE 4

Camera

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ image as projection of 3D world: leads to loss of depth data ◮ estimate depths through known size of an object and size of

the object in the image.

◮ error-prone, even in human visual perception ◮ not applicable outside of known surroundings ◮ passive approach

Stump in Sequoia National Park. [1, p. 529, fig. 2]

  • G. Glaser – Visual Perception Sensors

4 / 27

slide-5
SLIDE 5

Triangulation Approaches

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ compute point through known distance and measured angles

  • Triangulation. [7, p. 19, fig. 1]

Triangulation Calculation. [7, p. 20, fig. 1]

  • G. Glaser – Visual Perception Sensors

5 / 27

slide-6
SLIDE 6

Stereoscopic Cameras

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ one camera not sufficient for meaningful depth measurements ◮ use second camera to recover lost dimension ◮ triangulate distance

◮ known baseline between cameras ◮ corresponding points ◮ measured angles

◮ passive approach

Rotated stereo-camera rig and a Kinect. [6, p. 5, fig. 1.2]

  • G. Glaser – Visual Perception Sensors

6 / 27

slide-7
SLIDE 7

Stereoscopic Cameras

Problems

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ identification of

corresponding points in both images

◮ occlusion ◮ computationally expensive ◮ depends on illumination ◮ cameras need to be

synchronized

Stereo-Camera example. [5, p. 38, fig. 2.1]

  • G. Glaser – Visual Perception Sensors

7 / 27

slide-8
SLIDE 8

Structured-Light

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ project additional information on the object to allow recovery

  • f lost depth dimension

◮ several different approaches

◮ time multiplexing ◮ spatial multiplexing ◮ wavelength multiplexing

  • G. Glaser – Visual Perception Sensors

8 / 27

slide-9
SLIDE 9

Binary Projection

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ one camera, one projector ◮ several passes required ◮ deformity of lines as measure for depth ◮ time multiplexing ◮ active approach

Binary projection. [7, p. 30, fig. 1] Binary projection at different times t. [7, p. 33, fig. 1]

  • G. Glaser – Visual Perception Sensors

9 / 27

slide-10
SLIDE 10

Binary Projection

Problems

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ frames taken at different points in time

◮ time multiplexing ◮ not applicable for moving objects

◮ points directly on edges are uncertain

◮ soultion: gray code pattern Gray code projection at dif- ferent times t. [7, p. 33, fig. 1]

  • G. Glaser – Visual Perception Sensors

10 / 27

slide-11
SLIDE 11

Microsoft Kinect

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ RGB camera: 30fps @

640x480px

◮ spatial multiplexing ◮ USB 2.0 ◮ depth image: 30fps @

320x240px

◮ practical range [4]

◮ 0.8-4.5m in default mode ◮ 0.4-3m in near mode Microsoft Kinect. [3, p. 2, fig. 1-1]

  • G. Glaser – Visual Perception Sensors

11 / 27

slide-12
SLIDE 12

Microsoft Kinect

IR Laser Emitter

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ projection

◮ pseudo random noise-like

pattern

◮ 830nm wavelength

◮ laser

◮ heated/cooled to maintain

wavelength

◮ 70mW output power ◮ eye safety through scattering Projected IR pattern. [3, p. 12, fig. 2-2]

  • G. Glaser – Visual Perception Sensors

12 / 27

slide-13
SLIDE 13

Microsoft Kinect

Depth Image

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ IR camera image compared to known pattern

◮ disturbances can be used to calculate distances

◮ distances visualized as depth images

◮ red areas: close ◮ blue areas: further away ◮ black areas: no depth information available Depth image and corresponding RGB image. [3, p. 9, fig. 1-3]

  • G. Glaser – Visual Perception Sensors

13 / 27

slide-14
SLIDE 14

Microsoft Kinect

Problems

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ overexposure of IR camera

◮ by sunlight (only usable indoors) ◮ by reflecting surfaces

◮ only close range distances

◮ limited by laser output

◮ translucent objects not measurable ◮ latency of ~100ms [4] ◮ active approach, not easy to scale-out

◮ interferences with projected patterns

  • G. Glaser – Visual Perception Sensors

14 / 27

slide-15
SLIDE 15

Triangulation Approaches Conclusion

Stereo Cameras

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ good to calculate depths for distinct markers

◮ otherwise computationally expensive

◮ works indoors and outdoors ◮ completely passive, scaling out is possible without problems

  • G. Glaser – Visual Perception Sensors

15 / 27

slide-16
SLIDE 16

Triangulation Approaches Conclusion

Structured-Light Cameras

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ all approaches

◮ trouble measuring reflecting or transparent objects

◮ time multiplexing

◮ depth calculation for whole field of vision ◮ only for stationary objects

◮ spatial multiplexing (Kinect)

◮ computation done by hardware ◮ pretty complete depth map ◮ occluded areas ◮ too close or too far points

◮ wavelength multiplexing

◮ depth calculation with one photo ◮ low spatial resolution achievable

  • G. Glaser – Visual Perception Sensors

16 / 27

slide-17
SLIDE 17

Time of Flight Approaches

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ actively send out a signal ◮ measure time until reflection returns ◮ Light: P = 299.792.458 m

s ∗t

2

Simple ToF measurement. [8, p. 28, fig. 1.14]

  • G. Glaser – Visual Perception Sensors

17 / 27

slide-18
SLIDE 18

Depth Camera

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ active approach ◮ TX: illuminates whole scene

with array of IR emitters

◮ RX: ToF-receiver grid ◮ commonly used: sinus

modulation for emitted light

◮ measure point in time when

emitted signal returns

◮ calculate distance through

ToF

MESA Imaging SR4000, IR emitters. [8, p. 32, fig. 1.16]

  • G. Glaser – Visual Perception Sensors

18 / 27

slide-19
SLIDE 19

Depth Camera

Problems

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ hardware restrictions

◮ IR-emitter and

ToF-receievers in different position

◮ simulate central emitter to

avoid occlusion effects

◮ falsification of measurements

through multi path hopping

◮ point B will measure a

combination of two distances

◮ accurate time measurement

required

Pattern of IR emitters to avoid occlusion. [8, p. 34, fig. 1.17] Multipath phenomenon. [8, p. 104, fig. 3.16]

  • G. Glaser – Visual Perception Sensors

19 / 27

slide-20
SLIDE 20

Kinect V2

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ depth image: 50fps @

512x424px

◮ range 0.5-8m [4] ◮ latency of ~50ms [4] ◮ square wave modulation ◮ differential pixel array

◮ switches with square wave ◮ save returned light ◮ difference used to compute

distances

◮ high volume of data, requires

USB 3.0

Kinect V2. [4, p. 6, fig. 1-5]

  • G. Glaser – Visual Perception Sensors

20 / 27

slide-21
SLIDE 21

LIDAR

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ Light Detection And Ranging ◮ sends out single laser beam ◮ ToF to calculate distance ◮ single point sampling

◮ mirrors rotate laser beam to

scan line of points

◮ additional rotation possible

to scan area instead of line

Simple ToF measurement. [8, p. 28, fig. 1.14] Point clouds created by rotated line scanners. [2, p. 46, fig. 2.21]

  • G. Glaser – Visual Perception Sensors

21 / 27

slide-22
SLIDE 22

LIDAR

Problems

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ loss of spatial resolution with increased measurement distance ◮ transparent objects can not be measured ◮ mechanical moving parts

  • G. Glaser – Visual Perception Sensors

22 / 27

slide-23
SLIDE 23

Time-of-Flight

Conclusion

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ high laser outputs possible

◮ high measurement range ◮ sunlight can be compensated

◮ high sampling rates possible ◮ dynamic measurement range

◮ short and long distances can be measured together

  • G. Glaser – Visual Perception Sensors

23 / 27

slide-24
SLIDE 24

Conclusion

Required Ambient Lighting

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ structured light approaches require dark surroundings

◮ often used for optical measurements and inspection in industrial

robotics

◮ very precise measurements

◮ LIDAR can be built for outdoor usage ◮ other active approaches falsified/annulled by direct sunlight

  • G. Glaser – Visual Perception Sensors

24 / 27

slide-25
SLIDE 25

Conclusion

Computational Costs

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ active approaches

◮ distance calculation mostly handled by hardware

◮ stereoscopic cameras

◮ expensive: calculate matching points in both images

  • G. Glaser – Visual Perception Sensors

25 / 27

slide-26
SLIDE 26

Conclusion

Moving Objects

Motivation Triangulation Approaches Time of Flight Approaches Conclusion References

◮ depth cameras, spatial multiplexing structured light

◮ well suited ◮ record whole scene at single point in time

◮ binary projection

◮ not usable, time encoding through different frames

◮ LIDAR

◮ suitability depends on sampling rate and object movement Triangulation Time-of-Flight Stereo Camera Binary Projection Kinect Kinect V2 LIDAR

  • utdoor usability

✓ ✗ ✗ ✗ (✓) complete depth map (✗) ✓ ✓ ✓ (✓) passive ✓ ✗ ✗ ✗ ✗ scale out ✓ (✗) (✗) (✓) (✓) moving parts ✗ ✗ ✗ ✗ ✓ “cheap” (✓) ✗ ✓ ✓ ✗

  • G. Glaser – Visual Perception Sensors

26 / 27

slide-27
SLIDE 27

[1]

  • Prof. Dr. Leonie Dreschler-Fischer. Interactive visual computing lecture ws2012. Handouts 2012.

[2] Joachim Hertzberg, Kai Lingemann, and Andreas Nüchter. Mobile Roboter. Springer, Berlin, 2009. ISBN 978-3-642-01726-1. [3] Jeff Kramer, Matt Parker, Daniel Herrera C., Nicolas Burrus, and Florian Echtler. Hacking the Kinect. Apress, New York, 2012. ISBN 978-1-4302-3868-3. [4] Mansib Rahman. Beginning Microsoft Kinect for Windows SDK 2.0. Apress, Berkeley, CA, 2017. ISBN 978-1-4842-2316-1. Motion and Depth Sensing for Natural User Interfaces. [5] Stephen Se and Nick Pears. Passive 3D Imaging, pages 35–94. Springer London, London, 2012. ISBN 978-1-4471-4063-4. DOI:10.1007/978-1-4471-4063-4_2. URL https://doi.org/10.1007/978-1-4471-4063-4_2. [6] Jan Smisek, Michal Jancosek, and Tomas Pajdla. 3D with Kinect, pages 3–25. Springer London, London,

  • 2013. ISBN 978-1-4471-4640-7. DOI:10.1007/978-1-4471-4640-7_1. URL

https://doi.org/10.1007/978-1-4471-4640-7_1. [7]

  • Prof. Dr. Frank Steinicke. User interface & software & technology lecture ws2014. Chapter 5: 3D Capturing

Systems (Slides by Susanne Schmidt), 2014. [8] Pietro Zanuttigh, Giulio Marin, Carlo Dal Mutto, Fabio Dominio, Ludovico Minto, and Guido Maria Cortelazzo. Time-of-flight and structured light depth cameras : technology and applications. Springer, Switzerland, 2016. ISBN 978-3-319-30973-6.

  • G. Glaser – Visual Perception Sensors

27 / 27