Gesture recognition for Smartphones/Wearables Gestures hands, face, - - PowerPoint PPT Presentation

gesture recognition for smartphones wearables gestures
SMART_READER_LITE
LIVE PREVIEW

Gesture recognition for Smartphones/Wearables Gestures hands, face, - - PowerPoint PPT Presentation

Margarita Grinvald Gesture recognition for Smartphones/Wearables Gestures hands, face, body movements non-verbal communication human interaction 2 Gesture recognition interface with computers increase usability intuitive


slide-1
SLIDE 1

Margarita Grinvald

Gesture recognition for Smartphones/Wearables

slide-2
SLIDE 2

2

Gestures

▪ hands, face, body movements ▪ non-verbal communication ▪ human interaction

slide-3
SLIDE 3

3

Gesture recognition

▪ interface with computers ▪ increase usability ▪ intuitive interaction

slide-4
SLIDE 4

4

▪ Contact type: ▪ Touch based ▪ Non-contact type: ▪ Device gesture ▪ Vision based ▪ Electrical Field Sensing (EFS)

Gesture sensing

slide-5
SLIDE 5

▪ miniaturisation ▪ lack tactile clues ▪ no link between physical and digital interactions ▪ computational power

5

Issues on mobile devices

slide-6
SLIDE 6

▪ augment environment with digital information

6

Approaches

Sixthsense [Mistry et al. SIGGRAPH 2009] Skinput [Harrison et al. CHI 2010] OmniTouch [Harrison et al. UIST 2011]

slide-7
SLIDE 7

7

Approaches

▪ augment hardware

In-air typing interface for mobile devices with vibration feedback [Niikura et al. SIGGRAPH 2010] A low-cost transparent electric field sensor for 3D interaction [Le Goc et al. CHI 2014] MagGetz [Hwang et al. UIST 2013]

slide-8
SLIDE 8

8

▪ combine devices

Approaches

▪ efficient algorithms

In-air gestures around unmodified mobile devices [Song et al. UIST 2014] Duet: Exploring Joint interactions on a smart phone and a smart watch [Chen et al. CHI 2014]

slide-9
SLIDE 9

▪ augment environment with visual information ▪ interact through natural hand gestures ▪ wearable to be truly mobile

9

Sixthsense [Mistry et al. SIGGRAPH 2009]

slide-10
SLIDE 10

10

Color markers Camera Projector Mirror Smartphone

slide-11
SLIDE 11

11

Support for arbitrary surfaces

slide-12
SLIDE 12

12

Support for multitouch

slide-13
SLIDE 13

13

Limitations

▪ inability track surfaces ▪ differentiate hover and click ▪ accuracy limitations

slide-14
SLIDE 14

▪ skin as input canvas ▪ wearable bio-acoustic sensor ▪ localisation of finger tap

14

Skinput [Harrison et al. CHI 2010]

slide-15
SLIDE 15

15

Projector Armband

slide-16
SLIDE 16

16

Mechanical phenomena

▪ finger tap on skin generates acoustic energy ▪ some energy becomes sound waves ▪ some energy transmitted through the arm

slide-17
SLIDE 17

17

slide-18
SLIDE 18

18

Transverse waves

slide-19
SLIDE 19

19

Longitudinal waves

slide-20
SLIDE 20

▪ array of tuned vibrations sensors ▪ sensitive only to motion perpendicular to skin ▪ two sensing arrays to disambiguate different armband positions.

20

Sensing

slide-21
SLIDE 21

21

Sensor packages Weights

slide-22
SLIDE 22

▪ sensor data segmented into taps ▪ ML classification of location ▪ initial training stage

22

Tap localisation

slide-23
SLIDE 23

23

slide-24
SLIDE 24

24

▪ lack of support of other surfaces than skin ▪ no multitouch support ▪ no touch drag movement

Limitations

slide-25
SLIDE 25

25

▪ appropriate on demand ad hoc surfaces ▪ depth sensing and projection wearable ▪ depth driven template matching

OmniTouch [Harrison et al. UIST 2011]

slide-26
SLIDE 26

26

Depth Camera Projector

slide-27
SLIDE 27

▪ multitouch finger tracking on arbitrary surfaces ▪ no calibration or training ▪ resolve position and distinguish hover from click

27

Finger tracking

slide-28
SLIDE 28

28

Finger segmentation

Depth map Depth map gradient

slide-29
SLIDE 29

29

Finger segmentation

Candidates Tip estimation

slide-30
SLIDE 30

30

Click detection

Finger hovering Finger clicking

slide-31
SLIDE 31

▪ expand application space with graphical feedback ▪ track surface on which rendered ▪ update interface as surface moves

31

On demand interfaces

slide-32
SLIDE 32

32

Interface ‘glued’ to surface

slide-33
SLIDE 33

33

slide-34
SLIDE 34

▪ vision based 3D input interface ▪ detect keystroke action in the air ▪ provide vibration feedback

34

In-air typing interface for mobile devices with vibration feedback

[Niikura et al. SIGGRAPH 2010]

slide-35
SLIDE 35

35

Camera white LEDs vibration motor

slide-36
SLIDE 36

▪ high frame rate camera ▪ wide angle lens needs distortion correction ▪ skin colour extraction to detect fingertip ▪ estimate fingertip translation, rotation and scale

36

Tracking

slide-37
SLIDE 37

▪ difference of the dominant frequency of the fingertips scale to detect keystroke ▪ tactile feedback is important ▪ vibration feedback is conveyed after a keystroke

37

Keystroke feedback

slide-38
SLIDE 38

38

▪ camera is rich and flexible but with limitations ▪ minimal distance between sensor and scene ▪ sensitivity to lighting changes ▪ computational overheads ▪ high power requirements

Vision limitations

slide-39
SLIDE 39

39

▪ smartphone augmented with EFS ▪ resilient to illumination changes ▪ mapping measurements to 3D finger positions.

A low-cost transparent electric field sensor for 3D interaction

[Le Goc et al. CHI 2014]

slide-40
SLIDE 40

40

Drive electronics Electrode array

slide-41
SLIDE 41

▪ microchip built-in 3D positioning has low accuracy ▪ Random Decision Forests for regression on raw signal data ▪ speed and accuracy

41

Recognition

slide-42
SLIDE 42

42

slide-43
SLIDE 43

▪ tangible control widgets for richer tactile clues ▪ wider interaction area ▪ low cost and user configurable unpowered magnets

43

MagGetz [Hwang et al. UIST 2013]

slide-44
SLIDE 44

44

Magnetic fields Tangibles

slide-45
SLIDE 45

▪ traditional physical input controls with magnets ▪ magnetic traces change on widget state change ▪ track physical movement of control widgets

45

Tangibles

slide-46
SLIDE 46

46

Tangibles magnetism

Toggle switch Slider

slide-47
SLIDE 47

47

slide-48
SLIDE 48

▪ object damage by magnets ▪ magnetometer limitations

48

Limitations

slide-49
SLIDE 49

▪ extend interaction space with gesturing ▪ mobile devices RGB camera ▪ robust ML based algorithm

49

In-air gestures around unmodified mobile devices

[Song et al. UIST 2014]

slide-50
SLIDE 50

▪ detection of salient hand parts (fingertips) ▪ works without relying on highly discriminative depth data and rich computational resources ▪ no strong assumption about users environment ▪ reasonably robust to rotation and depth variation

50

Gesture recognition

slide-51
SLIDE 51

▪ real time algorithm ▪ pixel labelling with random forests ▪ techniques to reduce memory footprint of classifier

51

Recognition algorithm

slide-52
SLIDE 52

52

Recognition steps

RGB input Segmentation Labeling

slide-53
SLIDE 53

▪ division of labor ▪ works on many devices ▪ new apps enabled just by collecting new data

53

Applications

slide-54
SLIDE 54

54

slide-55
SLIDE 55

55

slide-56
SLIDE 56

▪ beyond usage of single device ▪ allow individual input and output ▪ joint interactions smart phone and smart watch

56

Duet: Exploring joint interactions on a smart phone and a smart watch

[Chen et al. CHI 2014]

slide-57
SLIDE 57

▪ conversational duet ▪ foreground interaction ▪ background interaction

57

Design space theory

slide-58
SLIDE 58

58

Design space

slide-59
SLIDE 59
slide-60
SLIDE 60

60

Design space

slide-61
SLIDE 61
slide-62
SLIDE 62

▪ ML techniques on accelerometer data ▪ handedness recognition ▪ promising accuracy

62

Gesture recognition

slide-63
SLIDE 63

▪ wearables extend interaction space to everyday surfaces ▪ augmented hardware in general provides an intuitive interface ▪ no additional hardware is preferable but there are still computational limitations ▪ combination of devices may be redundant

63

Summary

slide-64
SLIDE 64

▪ SixthSense: a wearable gestural interface [Mistry et al. SIGGRAPH 2009] ▪ Skinput: Appropriating the Body As an Input Surface [Harrison et al. CHI 2010] ▪ OmniTouch: Wearable Multitouch Interaction Everywhere [Harrison et al. UIST 2011] ▪ In-air typing interface for mobile devices with vibration feedback [Niikura et al. SIGGRAPH 2010] ▪ A Low-cost Transparent EF Sensor for 3D Interaction on Mobile Devices [Le Goc et al. CHI 2014] ▪ MagGetz: customizable passive tangible controllers on and around [Hwang et al. UIST 2013] ▪ In-air gestures around unmodified mobile devices mobile devices [Song et al. UIST 2014] ▪ Duet: Exploring Joint Interactions on a Smart Phone and a Smart Watch [Chen et al. CHI 2014]

64

References