ON HUMAN ACTION Volker Krger Dept. of Mechanical and Production - - PowerPoint PPT Presentation

on human action
SMART_READER_LITE
LIVE PREVIEW

ON HUMAN ACTION Volker Krger Dept. of Mechanical and Production - - PowerPoint PPT Presentation

ON HUMAN ACTION Volker Krger Dept. of Mechanical and Production Engineering Aalborg University vok@m-tech.aau.dk Saturday, April 21, 2012 () Cleaning the Kitchen Observation World Model Reproduction/Recognition Saturday, April 21, 2012


slide-1
SLIDE 1

ON HUMAN ACTION

Volker Krüger

  • Dept. of Mechanical and Production Engineering

Aalborg University vok@m-tech.aau.dk

Saturday, April 21, 2012 ()

slide-2
SLIDE 2

Observation World Model Reproduction/Recognition Cleaning the Kitchen

Saturday, April 21, 2012 ()

slide-3
SLIDE 3

Observation World Model Reproduction/Recognition

Cleaning the Kitchen

What is the action?

Saturday, April 21, 2012 ()

slide-4
SLIDE 4

Observation World Model Reproduction/Recognition

Cleaning the Kitchen

What is the action? Grasping a plate?

Saturday, April 21, 2012 ()

slide-5
SLIDE 5

Observation World Model Reproduction/Recognition

Cleaning the Kitchen

What is the action? Grasping a plate? Putting plates upright?

Saturday, April 21, 2012 ()

slide-6
SLIDE 6

Observation World Model Reproduction/Recognition

Cleaning the Kitchen

What is the action? Grasping a plate? Putting plates upright? Removing plates from the table?

Saturday, April 21, 2012 ()

slide-7
SLIDE 7

Observation World Model Reproduction/Recognition

Cleaning the Kitchen

What is the action? Grasping a plate? Putting plates upright? Removing plates from the table? Filling the dish washer?

Saturday, April 21, 2012 ()

slide-8
SLIDE 8

Observation World Model Reproduction/Recognition

Cleaning the Kitchen

What is the action? Grasping a plate? Putting plates upright? Removing plates from the table? Filling the dish washer? Cleaning the kitchen?

Saturday, April 21, 2012 ()

slide-9
SLIDE 9

Observation World Model Reproduction/Recognition

Cleaning the Kitchen

What is the action? Grasping a plate? Putting plates upright? Removing plates from the table? Filling the dish washer? Cleaning the kitchen? So what does it mean to understand the meaning of an action?

Saturday, April 21, 2012 ()

slide-10
SLIDE 10

Observation World Model Reproduction/Recognition

Cleaning the Kitchen

The meaning of an action is the state change that the physical movement

  • f an actor causes to the world state

space. That can be on different levels of

  • abstraction. At least, this is the goal.

Saturday, April 21, 2012 ()

slide-11
SLIDE 11
  • What is the person doing?
  • Li Fei-Fei, CVPR10

Saturday, April 21, 2012 ()

slide-12
SLIDE 12
  • What is the person doing?
  • Objects and actions are intertwined
  • Li Fei-Fei, CVPR10

Saturday, April 21, 2012 ()

slide-13
SLIDE 13
  • What is the person doing?
  • Objects and actions are intertwined
  • Objects prime actions, actions prime objects
  • Li Fei-Fei, CVPR10

Saturday, April 21, 2012 ()

slide-14
SLIDE 14
  • Gibson, J.J. (1977). The theory of affordances. In R. Shaw & J. Bransford (eds.), Perceiving, Acting and Knowing. Hillsdale, NJ: Erlbaum.
  • Norman, D. (1988). The Psychology of Everyday Things. New

York, Basic Books, pp. 87-92.

  • Humphreys, G. et al. The interaction of attention and action: From seeing action to acting on perception. British Journal of Psychology

(2010), 101, 185–206

The world is perceived not only in terms of object shapes and spatial relationships but also in terms of object possibilities for action (affordances). perception drives action.

Saturday, April 21, 2012 ()

slide-15
SLIDE 15
  • Gallese et al. ”Action Recognition in the premotor cortex”, Brain, vol. 119, no. 2, 1996.
  • Nishitani et al. ”Broca’s Region: From Action to Language” Physiology, vol. 20, 2005.
  • Rizzolatti et al. ”Neurophysiological Mechanisms Underlying the Unterstanding and Imitation of Action” Nature Reviews, vol

2, 2001.

  • Newtson: “The Objective Basis of Behavior Units”, Journal of Personality and Social Psychology, vol 35(12), 1977.

perception and action share the same symbolic structure

Saturday, April 21, 2012 ()

slide-16
SLIDE 16
  • Gallese et al. ”Action Recognition in the premotor cortex”, Brain, vol. 119, no. 2, 1996.
  • Nishitani et al. ”Broca’s Region: From Action to Language” Physiology, vol. 20, 2005.
  • Rizzolatti et al. ”Neurophysiological Mechanisms Underlying the Unterstanding and Imitation of Action” Nature Reviews, vol

2, 2001.

  • Newtson: “The Objective Basis of Behavior Units”, Journal of Personality and Social Psychology, vol 35(12), 1977.

perception and action share the same symbolic structure spoken language and visible movements use same cognitive substrate

Saturday, April 21, 2012 ()

slide-17
SLIDE 17
  • Objects and Actions are inseparably intertwined.

OBJECT ACTION COMPLEXES (OACS)

Saturday, April 21, 2012 ()

slide-18
SLIDE 18
  • Objects and Actions are inseparably intertwined.
  • Categories are determined (and also limited) by the action an agent can perform and by

the attributes of the world it can perceive;

OBJECT ACTION COMPLEXES (OACS)

Saturday, April 21, 2012 ()

slide-19
SLIDE 19
  • Objects and Actions are inseparably intertwined.
  • Categories are determined (and also limited) by the action an agent can perform and by

the attributes of the world it can perceive;

  • Entities “things” in the world of a robot (or a human) will only become semantically

useful “objects” through the action that the agent can/will perform on them.

OBJECT ACTION COMPLEXES (OACS)

Saturday, April 21, 2012 ()

slide-20
SLIDE 20
  • OAC

“Thing” & Movement “Thing” & changed Objects & Actions have arisen & Attributes changed Attributes have arisen if this process is successful Success ? Code-similarity M d i t Human Neuronal “code” Code similarity emerges only through the fact that both codes Consistency with world Measured against: Robot “code” code describe the same physical entity Consistency with world Novelty, Drives, etc. OACs are code-independent

  • Saturday, April 21, 2012 ()
slide-21
SLIDE 21
  • Objects

OAC

“movement” & Thing “movement” & changed Objects & Actions have arisen & Attributes changed Attributes have arisen if this process is successful Success ? Code-similarity Measured against: Human Neuronal “code” y emerges only through the fact that both codes Consistency with world Measured against: Robot “code” code describe the same physical entity Consistency with world Novelty, Drives, etc. OACs are code-independent

  • Saturday, April 21, 2012 ()
slide-22
SLIDE 22
  • l

Object Perspective

G i Pen Pi h G i

Action Perspective

Container (full) tilting Container (empty) Grip Pinch Grip Success? Success?

Consistency w. world Novelty, Drives, etc.

Grip Can Power Grip

Consistency w. world Novelty, Drives, etc.

Success?

Consistency w. world Novelty, Drives, etc.

Task Perspective

“drink” C “put on table” Grip Cup Handle Grip drink Grip Cup From-top Grip Success?

Consistency w. world Novelty, Drives, etc.

Success?

Consistency w. world Novelty, Drives, etc.

  • Saturday, April 21, 2012 ()
slide-23
SLIDE 23
  • Object Action Complexes (OACs)
  • Actions define the meaning of Objects
  • Objects suggest Actions (affordance)

Saturday, April 21, 2012 ()

slide-24
SLIDE 24
  • Object Action Complexes (OACs)
  • Actions define the meaning of Objects
  • Objects suggest Actions (affordance)
  • OACs are associations of objects and affordances
  • Affordances can be expressed by STRIPS like-rules

Saturday, April 21, 2012 ()

slide-25
SLIDE 25
  • Object Action Complexes (OACs)
  • Actions define the meaning of Objects
  • Objects suggest Actions (affordance)
  • OACs are associations of objects and affordances
  • Affordances can be expressed by STRIPS like-rules
  • Associative memory ensures that
  • Object representations (and other preconditions) evoke

affordances

  • Representations of affordances (and other preconditions) evoke objects

Saturday, April 21, 2012 ()

slide-26
SLIDE 26
  • Affordances are “unidirectional”: Objects affords actions
  • OACs are “bidirectional”: Object affords actions Actions suggest objects
  • OACs can be chained (new complex OACs from simpler OACs “Tasks from skills =

Planning”)

OACS VS. AFFORDANCES

Saturday, April 21, 2012 ()

slide-27
SLIDE 27
  • action hierarchy
  • Actions involve objects, Movements do not
  • Action primitives are the atomic entities
  • vital due to computational / combinatorial aspects

Activities Actions Action Primitives Movements

ACTION PRIMITIVES WITHIN OACS

Saturday, April 21, 2012 ()

slide-28
SLIDE 28
  • action hierarchy
  • Actions involve objects, Movements do not
  • Action primitives are the atomic entities
  • vital due to computational / combinatorial aspects

Activities Actions Action Primitives Movements

action primitives are atomic building blocks of actions. They

ACTION PRIMITIVES WITHIN OACS

Saturday, April 21, 2012 ()

slide-29
SLIDE 29
  • action hierarchy
  • Actions involve objects, Movements do not
  • Action primitives are the atomic entities
  • vital due to computational / combinatorial aspects

Activities Actions Action Primitives Movements

action primitives are atomic building blocks of actions. They

  • are meant to change the world state in a specific manner

ACTION PRIMITIVES WITHIN OACS

Saturday, April 21, 2012 ()

slide-30
SLIDE 30
  • action hierarchy
  • Actions involve objects, Movements do not
  • Action primitives are the atomic entities
  • vital due to computational / combinatorial aspects

Activities Actions Action Primitives Movements

action primitives are atomic building blocks of actions. They

  • are meant to change the world state in a specific manner
  • require a certain world state

ACTION PRIMITIVES WITHIN OACS

Saturday, April 21, 2012 ()

slide-31
SLIDE 31
  • action hierarchy
  • Actions involve objects, Movements do not
  • Action primitives are the atomic entities
  • vital due to computational / combinatorial aspects

Activities Actions Action Primitives Movements

action primitives are atomic building blocks of actions. They

  • are meant to change the world state in a specific manner
  • require a certain world state
  • can generate a specific change to world state

ACTION PRIMITIVES WITHIN OACS

Saturday, April 21, 2012 ()

slide-32
SLIDE 32
  • action hierarchy
  • Actions involve objects, Movements do not
  • Action primitives are the atomic entities
  • vital due to computational / combinatorial aspects

Activities Actions Action Primitives Movements

action primitives are atomic building blocks of actions. They

  • are meant to change the world state in a specific manner
  • require a certain world state
  • can generate a specific change to world state

OACS contain the sensing capabilities (visual, haptic, force torque)

ACTION PRIMITIVES WITHIN OACS

Saturday, April 21, 2012 ()

slide-33
SLIDE 33

Actions Objects Context

P(o, a, c)

Actions Objects

Parameters

P(o, a, w) ≡ P(a, w|o)

  • Li Fei-Fei, CVPR10

ATTEMPT TO IMPLEMENT OACS

Saturday, April 21, 2012 ()

slide-34
SLIDE 34
  • Fig. 2: The figure shows the different steps involved in learning

Saturday, April 21, 2012 ()

slide-35
SLIDE 35

UNSUPERVISED LEARNING OF WORLD STATE SPACE

  • Identify statistics in the effect space O
  • Propagate the clustering of the effect space to the human action space H

Use object info to segment O Find primitives Group primitives with same effect in the object space Segment and group primitives in action space in object space Object space Action space Input features

[H O]

Saturday, April 21, 2012 ()

slide-36
SLIDE 36
  • Parameterization is here object location + (speed and direction).
  • Unsupervised learning of context-free grammar
  • recursive construction of HMM
  • Dirichlet Process

Saturday, April 21, 2012 ()

slide-37
SLIDE 37
  • Dirichlet Processes generalize finite mixture models to infinite mixture models
  • choice of mixture number is data-driven, similar to k-means clustering
  • Dirichlet Process find the number of mixtures automatically.
  • DPs and HDPs are unsupervised.

Gaussian

c x ci xi N p θ

Slide partially borrowed from Teg Grenager

Saturday, April 21, 2012 ()

slide-38
SLIDE 38

ci xi N p θ α G0

Slide partially borrowed from Teg Grenager

Saturday, April 21, 2012 ()

slide-39
SLIDE 39

Slide partially borrowed from Teg Grenager

Saturday, April 21, 2012 ()

slide-40
SLIDE 40
  • What we have now is
  • states: clusters of trajectories that all have the same effect
  • detecting their grammatical relationship is trivial

Sanmohan, V. Krüger, D. Kragic, and H. Kjellström. Automatic Primitive Segmentation and Action Recognition. Advanced Robotics, 25(6-7):871– 891, 2011.

  • V. Krueger, Sanmohan, D. Herzog, A. Ude, and D. Kragic. Learning Actions from Observations. IEEE Robotics and Automation

Magazine, 17(2):30–43, 2010.

Saturday, April 21, 2012 ()

slide-41
SLIDE 41
  • What we have now is
  • states: clusters of trajectories that all have the same effect
  • detecting their grammatical relationship is trivial
  • Next step builds a model for the observed actions within each cluster.

Sanmohan, V. Krüger, D. Kragic, and H. Kjellström. Automatic Primitive Segmentation and Action Recognition. Advanced Robotics, 25(6-7):871– 891, 2011.

  • V. Krueger, Sanmohan, D. Herzog, A. Ude, and D. Kragic. Learning Actions from Observations. IEEE Robotics and Automation

Magazine, 17(2):30–43, 2010.

Saturday, April 21, 2012 ()

slide-42
SLIDE 42
  • What we have now is
  • states: clusters of trajectories that all have the same effect
  • detecting their grammatical relationship is trivial
  • Next step builds a model for the observed actions within each cluster.
  • Issues: right parameterization!! What matters?

Sanmohan, V. Krüger, D. Kragic, and H. Kjellström. Automatic Primitive Segmentation and Action Recognition. Advanced Robotics, 25(6-7):871– 891, 2011.

  • V. Krueger, Sanmohan, D. Herzog, A. Ude, and D. Kragic. Learning Actions from Observations. IEEE Robotics and Automation

Magazine, 17(2):30–43, 2010.

Saturday, April 21, 2012 ()

slide-43
SLIDE 43
  • Modeling clusters of trajectories
  • Parametric HMMs: Hidden Markov Models, that allow

for parametric means and covariances

  • Parameters have meaning
  • given by the object and the effects.

Wilson&Bobick, PAMI 99

The Fish was this big

PARAMETRIC HIDDEN MARKOV MODELS

Saturday, April 21, 2012 ()

slide-44
SLIDE 44
  • Tell me the action and the object,

and I know the movement (up to some uncertainty)!

  • Action and parameters infer joint settings

and pose: huge dimensionality reduction

  • Tracking is simplified, synthesis is trivial

On the fly demo, monocular(!!) data

MODELING ACTIONS IN OBJECT ACTION SPACE

Saturday, April 21, 2012 ()

slide-45
SLIDE 45
  • Action parameters in case of a table top scenario
  • : PHMM state, associated with a human pose
  • : object location on the table
  • : action identifier

w = (k, x, y) k x, y i

TRACKING IN OBJECT ACTION SPACE

Saturday, April 21, 2012 ()

slide-46
SLIDE 46
  • Action parameters in case of a table top scenario
  • : PHMM state, associated with a human pose
  • : object location on the table
  • : action identifier

w = (k, x, y) k x, y i

TRACKING IN OBJECT ACTION SPACE

to be estimated, can be constrained

Saturday, April 21, 2012 ()

slide-47
SLIDE 47
  • Action parameters in case of a table top scenario
  • : PHMM state, associated with a human pose
  • : object location on the table
  • : action identifier

w = (k, x, y) k x, y i

P(ωt, it|Z1 . . . Zt)) ≡ pt(ωt, it) = X

it−1

Z

ωt−1

pt(Zt|ωt, it)p(ωt, ii|ωt−1, it−1)pt−1(ωt−1, it−1)dωt−1

TRACKING IN OBJECT ACTION SPACE

to be estimated, can be constrained

  • Classical Bayesian Propagation over time

Saturday, April 21, 2012 ()

slide-48
SLIDE 48

RECOGNIZING ACTIONS

Saturday, April 21, 2012 ()

slide-49
SLIDE 49
  • Parametric action recognition
  • pointing, reaching, pushing and filling

actions.

  • parameters of the action are marginalized
  • ut

RECOGNIZING ACTIONS

Saturday, April 21, 2012 ()

slide-50
SLIDE 50

TRACKING IN OBJECT ACTION SPACE

Saturday, April 21, 2012 ()

slide-51
SLIDE 51
  • Monocular and multi-view

tracking

  • red dot marks the active camera
  • color of the ball is given by the

parameter uncertainty

TRACKING IN OBJECT ACTION SPACE

Saturday, April 21, 2012 ()

slide-52
SLIDE 52
  • 3x2 grid with 5 repetitions each.
  • Integrated error along the trajectory

RECOVERY OF PARAMETERS

✏ = v u u t Z

7

X

i=1

(fi(↵(t)) − ¯ fi(¯ ↵(t)))2 7 dt Z ↵(d)dt

Error recovered params

1 2 3 4 5 6 7 1 2 3 4 5 2 3 4

80cm 30cm error [cm]

Same error as for human ground-truth trajectories(!) Deviation Trajectory

1 2 3 4 5 6 7 1 2 3 4 5 2 3 4

80cm 30cm error [cm]

Saturday, April 21, 2012 ()

slide-53
SLIDE 53
  • Using more complex grammars
  • Pick and Place actions
  • Tracker switches between different action

primitives

A C B

10 20 30 40 50 60 70 80 90 100 −0.5 0.5 1 1.5

frame posterior probability Camera 2

action A action B

TRACKING IN OBJECT ACTION SPACE

  • Dennis Herzog and Volker Krueger. Tracking in Action Space. In

Human Motion: Understanding, Modeling, Capture and Animation, Workshop at ECCV 2010, 2010. Springer.

  • Dennis Herzog and Volker Krueger. Tracking in Action Space. Int.

Journal Computer Vision and Image Understanding (CVIU). submitted

Saturday, April 21, 2012 ()

slide-54
SLIDE 54
  • HOAP3 robot
  • arm movements are defined by

PHMMs

  • robot picks and places the objects

OBJECT-ACTION SPACE FOR ROBOT CONTROL

  • V. Krueger, Sanmohan, D. Herzog, A. Ude, and D. Kragic. Learning Actions from Observations. IEEE Robotics and Automation

Magazine, 17(2):30–43, 2010.

  • D. Herzog, A. Ude, and V. Krueger. Motion Imitation and Recognition using Parametric Hidden Markov Models. In Humanoids,

IEEE-RAS International Conference on Humanoid Robots, Daejeon, Korea, South, December 1-3, 2008

Saturday, April 21, 2012 ()

slide-55
SLIDE 55

LEARNING OF ACTION PRIMITIVES

  • Action primitives for motor control
  • Starting point: Khansari-Zadeh and Billard, Imitation

Learning of Globally Stable non-linear Point-to-Point Robot Motions using Non-linear Programming, IROS2010

  • SEDS-approach (see Billard’s and Calinon’s

presentation)

Target Demonstrations Reproductions Initial points −200 −100 −200 −100 100 200 300 400

ξ1(mm) ξ2(mm) ξ3(mm)

−200 200 −100 100 200 −400 −300 −200 −100 100 200

˙ ξ1(mm/s) ˙ ξ2(mm/s) ˙ ξ3(mm/s)

Saturday, April 21, 2012 ()

slide-56
SLIDE 56

NON-LINEAR DYNAMIC MODEL SETTING UP THE PROBLEM

Saturday, April 21, 2012 ()

slide-57
SLIDE 57

NON-LINEAR DYNAMIC MODEL SETTING UP THE PROBLEM

  • We model the movements using a dynamic model:
  • = Model parameters
  • = state vector

θ ξ ˙ ⇠ = f(⇠; ✓) + ✏

Saturday, April 21, 2012 ()

slide-58
SLIDE 58

NON-LINEAR DYNAMIC MODEL SETTING UP THE PROBLEM

  • We model the movements using a dynamic model:
  • = Model parameters
  • = state vector
  • To model , a Gaussian mixture model is used:

θ ξ ˙ ⇠ = f(⇠; ✓) + ✏ P(ξt,n, ˙ ξt,n) =

K

X

k=1

πkN(ξt,n, ˙ ξt,n; θk) f

Saturday, April 21, 2012 ()

slide-59
SLIDE 59

NON-LINEAR DYNAMIC MODEL SETTING UP THE PROBLEM

  • We model the movements using a dynamic model:
  • = Model parameters
  • = state vector
  • To model , a Gaussian mixture model is used:
  • where

θ ξ ˙ ⇠ = f(⇠; ✓) + ✏ P(ξt,n, ˙ ξt,n) =

K

X

k=1

πkN(ξt,n, ˙ ξt,n; θk)

covariance matrices θk = {πk, µk, Σk} covariance matrix of

µk =

  • µk

ξ

µk

˙ ξ

  • ,

Σk =

  • Σk

ξ

Σk

ξ ˙ ξ

Σk

˙ ξξ

Σk

˙ ξ

  • f

Saturday, April 21, 2012 ()

slide-60
SLIDE 60

NON-LINEAR DYNAMIC MODEL SETTING UP THE PROBLEM

  • We model the movements using a dynamic model:
  • = Model parameters
  • = state vector
  • To model , a Gaussian mixture model is used:
  • where
  • This can then be rewritten as

θ ξ ˙ ⇠ = f(⇠; ✓) + ✏ P(ξt,n, ˙ ξt,n) =

K

X

k=1

πkN(ξt,n, ˙ ξt,n; θk)

covariance matrices θk = {πk, µk, Σk} covariance matrix of

µk =

  • µk

ξ

µk

˙ ξ

  • ,

Σk =

  • Σk

ξ

Σk

ξ ˙ ξ

Σk

˙ ξξ

Σk

˙ ξ

  • ˆ

˙ ξ =

K

  • k=1

πkN(ξ; θk) K

i=1 πiN(ξ; θi)

(µk

˙ ξ + Σk ˙ ξξ(Σk ξ)−1(ξ − µk ξ))

f

Saturday, April 21, 2012 ()

slide-61
SLIDE 61

BUT WHAT ABOUT K?

  • Someone needs to decide!

−2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 x 10

8

−1 −0.5 0.5 1 1.5 2 x 10

8

Saturday, April 21, 2012 ()

slide-62
SLIDE 62

BUT WHAT ABOUT K?

  • Someone needs to decide!

−2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 x 10

8

−1 −0.5 0.5 1 1.5 2 x 10

8

  • We know that finding K is a principle

problem!

  • so that is fine.

Saturday, April 21, 2012 ()

slide-63
SLIDE 63

BUT WHAT ABOUT K?

  • Someone needs to decide!

−2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 x 10

8

−1 −0.5 0.5 1 1.5 2 x 10

8

  • We know that finding K is a principle

problem!

  • so that is fine.
  • But the big problem is this:
  • the actions to be learned must be known

in advance!! That is obviously a problem!

Saturday, April 21, 2012 ()

slide-64
SLIDE 64

BUT WHAT ABOUT K?

  • Someone needs to decide!

−2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 x 10

8

−1 −0.5 0.5 1 1.5 2 x 10

8

  • We know that finding K is a principle

problem!

  • so that is fine.
  • But the big problem is this:
  • the actions to be learned must be known

in advance!! That is obviously a problem!

  • We use Dirichlet process to find K.

Saturday, April 21, 2012 ()

slide-65
SLIDE 65

BUT WHAT ABOUT K?

  • Someone needs to decide!

−2.5 −2 −1.5 −1 −0.5 0.5 1 1.5 2 2.5 x 10

8

−1 −0.5 0.5 1 1.5 2 x 10

8

  • We know that finding K is a principle

problem!

  • so that is fine.
  • But the big problem is this:
  • the actions to be learned must be known

in advance!! That is obviously a problem!

  • We use Dirichlet process to find K.
  • Application of DPs here is non-trivial.

Saturday, April 21, 2012 ()

slide-66
SLIDE 66

RESULTS

  • Results for 3D movement “Letter N” captured with iCub: 4 Gaussians

Location Velocity

Saturday, April 21, 2012 ()

slide-67
SLIDE 67

WORKS GREAT!

  • Results for 3D movement “Letter S” captured with iCub: 4 Gaussians

Location Velocity

Saturday, April 21, 2012 ()

slide-68
SLIDE 68

WORKS GREAT!

  • Results for 3D movement “Letter C”: Comparison Training vs Simulation

Location Velocity

−400 −200 200 −300 −250 −200 −150 −100 −50 50 100 −100 100 200 300 400 500

ξ1(mm) ξ2(mm) ξ3(mm)

−100 −50 50 100 150 −50 50 100 150 −300 −250 −200 −150 −100 −50 50 100 150

˙ ξ1(mm/s) ˙ ξ2(mm/s) ˙ ξ3(mm/s)

Krüger et al. Imitation Learning of Non-Linear Point-to-Point Robot Motions using Dirichlet Processes. ICRA 2012

Saturday, April 21, 2012 ()

slide-69
SLIDE 69

SPIRIT OF OACS FOR INDUSTRIAL ROBOTS

Saturday, April 21, 2012 ()

slide-70
SLIDE 70

SKills are OACS for Industrial Applications

Saturday, April 21, 2012 ()

slide-71
SLIDE 71
  • pre- and post-conditions: Important for

robustness and planning

  • STRIPS-like planner
  • Markov Decision process

SKills are OACS for Industrial Applications

Saturday, April 21, 2012 ()

slide-72
SLIDE 72
  • pre- and post-conditions: Important for

robustness and planning

  • STRIPS-like planner
  • Markov Decision process

SKills are OACS for Industrial Applications Problem: Finding the right set of skills

Saturday, April 21, 2012 ()

slide-73
SLIDE 73

VOCABULARY OF TASKS

  • analyzed 566 tasks at Grundfos
  • task implementations
  • standard operation procedures

(SOPs)

Saturday, April 21, 2012 ()

slide-74
SLIDE 74

VOCABULARY OF SKILLS

Skill Description Move to To go from one location (station) to an-

  • ther in the factory

Locate To determine or specify the position of an

  • bject by searching or examining

Pick up To take hold of and lift up Place To arrange something in a certain spot or position Unload Unload a container: to remove, discharge

  • r empty the contents from a container

Shovel To take up and move objects with a shovel Check Quality control: the act or process of test- ing, verifying or examining Align To make an object come into precise adjustment or correct relative position to another object Open To move (as a door) from a closed position and make available for entry, passage or accessible Close To move (as a door) from an open position Press To press against with force in order to drive or impel. Release To let go or set free from restraint e.g. release a button Turn To turn a knob or handle

6 Transportation Skills 10 Assistive Skills

  • Move to <location>
  • Locate <object>
  • Pick up <object>
  • Place <object, coord>
  • Unload <container, coord>
  • Shovel <container, coordinate>
  • Pick up <object>
  • Place <object, coordinate>
  • Locate <object>
  • Press <object>
  • Check <object>
  • Align <object, object>
  • Open <object>
  • Close <object>
  • Release <object>
  • Turn <object>

Skill-complete with 13 skills

Saturday, April 21, 2012 ()

slide-75
SLIDE 75

PROGRAMING USING SKILLS

Saturday, April 21, 2012 ()

slide-76
SLIDE 76

SENSING IS THE KEY

Saturday, April 21, 2012 ()

slide-77
SLIDE 77

SENSING IS THE KEY

Grasping Skill(object) prior state

  • object is in field of view --> provides 3D coord
  • object is graspable (use 3D coord)
  • distance to object
  • grasping trajectory exists

execute grasping trajectory. Use force torque to already after the actual grasp verify for success posteriori:

  • object is in the gripper
  • location within the robot body space

Saturday, April 21, 2012 ()

slide-78
SLIDE 78

PLACE-SKILL(LOCATION)

  • prior
  • location in {table, shelf, magazin} (location given as bar code)
  • empty <location> is available and reachable
  • is the gripper free?
  • execute place skill + simultaneous verification using force torque
  • poterior
  • empty gripper
  • location not empty anymore
  • arm back in robot body space (note: breaching the robot body space: moving skills may not lead to a breach, but

manipulation skills may)

Saturday, April 21, 2012 ()

slide-79
SLIDE 79

IMPLEMENTATION LAYERS

  • Generalized Plans on different levels

Saturday, April 21, 2012 ()

slide-80
SLIDE 80

PROGRAMING USING SKILLS

  • programs can be generated automatically using a

planner

  • probabilistic using Markov decision process
  • binary using STRIPS planner

See the demo here: feeding Demo.mov - YouTube

Saturday, April 21, 2012 ()

slide-81
SLIDE 81

PROGRAMING USING SKILLS

  • programs can be generated automatically using a

planner

  • probabilistic using Markov decision process
  • binary using STRIPS planner
  • gripper: full/empty
  • magazin: filling level
  • robot location: discrete:
  • home
  • warehouse
  • feeder 1,2,3...
  • SLC full/empty
  • feeder: empty, apparently full, full

See the demo here: feeding Demo.mov - YouTube

Saturday, April 21, 2012 ()

slide-82
SLIDE 82
  • Object-Action Complexes
  • Unsupervised learning of action grammar based on effects of the observed actions
  • Modeling of human actions using SEDS-DMPs and PHMMs
  • only tested on simple scenarios, not clear how well it will scale
  • hand-generated “OCAs” / Skills for industrial scenario
  • Are OACs are good choice for industrial applications?
  • What about assembly tasks?
  • How should the skills be for collaboration? event-driven rather than effect-driven?

SUMMARY + CONCLUSIONS

Saturday, April 21, 2012 ()

slide-83
SLIDE 83

REFERENCES

  • The meaning of action
  • V. Krueger, D. Kragic, A. Ude, and C. Geib. Meaning of Action. Int. Journal on Advanced Robotics, Special issue on Imitative Robotics, T. Inamura and G.

Metta (eds.), 2007.

  • A. Bobick and V. Krüger. Visual Analysis of Humans, T. Moeslund ,A. Hilton ,V. Krü̈ger and L.Segal (Eds.), chapter On Human Action, pages 279–288. Springer,

2011.

  • Action Primitives
  • Danica Kragic Dana Kulic and Volker Krüger. Visual Analysis of Humans, In T. Moeslund, A. Hilton , V. Krüger and L.Segal (Eds.), chapter Learning Action

Primitives, pages 333–353. Springer, 2011.

  • Sanmohan, V. Krüger, D. Kragic, and H. Kjellström. Automatic Primitive Segmentation and Action Recognition. Advanced Robotics, 25(6-7):871– 891, 2011.
  • V. Krueger, Sanmohan, D. Herzog, A. Ude, and D. Kragic. Learning Actions from Observations. IEEE Robotics and Automation Magazine, 17(2):30–43, 2010.
  • Simon Bøgh, Oluf Nielsen, Mikkel Pedersen, Volker Krüger: Does your Robot have Skills?. 43rd International Symposium on Robotics (ISR)
  • Modeling, recognition and generation of Actions
  • Dennis Herzog and Volker Krueger. Tracking in Actionspace. In Human Motion: Understanding, Modeling, Capture and Animation, Workshop at ECCV

2010, 2010. Springer.

  • Dennis Herzog and Volker Krueger. Tracking in Object-Action Space. Int. Journal Computer Vision and Image Understanding (CVIU). submitted
  • D. Herzog, A. Ude, and V. Krueger. Motion Imitation and Recognition using Parametric Hidden Markov Models. In Humanoids, IEEE-RAS In- ternational

Conference on Humanoid Robots, Daejeon, Korea, South, December 1-3, 2008.

  • Volker Krüger, Vadim Tikhanoff, Lorenzo Natale and Giulio Sandini: Imitation Learning of Non-Linear Point-to-Point Robot Motions

using Dirichlet Processes IEEE 2012 Conf. of Robotics and Automation, ICRA 2012

  • Carsten Høilund, Volker Krüger and Thomas Moeslund. Evaluation of Human Body Tracking System for Gesture-based Programming of Industrial Robots

ICIEA2012

  • Systems
  • Carsten Høilund, Mikkel Pedersen and Volker Krüger. Using Human Gestures to Program Generic Skills for a Mobile Robot Arm in a Feeder Filling
  • Scenario. ICRA ECHORD-Workshop, 2012.

Saturday, April 21, 2012 ()

slide-84
SLIDE 84

THANKS

Questions? Comments? www.m-tech.aau.dk/Research+Groups/Robotics+and+Automation/

Saturday, April 21, 2012 ()