Visual Servoing, Intro Optimal Control Lecture 12 What will you - - PowerPoint PPT Presentation

visual servoing intro optimal control
SMART_READER_LITE
LIVE PREVIEW

Visual Servoing, Intro Optimal Control Lecture 12 What will you - - PowerPoint PPT Presentation

Visual Servoing, Intro Optimal Control Lecture 12 What will you take home today? Visual Servoing Interaction Matrix Control Law Case-Study: Learning-based approach Introduction Optimal Control Principle of Optimality Bellman Equation


slide-1
SLIDE 1

Visual Servoing, Intro Optimal Control

Lecture 12

slide-2
SLIDE 2

What will you take home today?

Visual Servoing Interaction Matrix Control Law Case-Study: Learning-based approach Introduction Optimal Control Principle of Optimality Bellman Equation Deriving LQR

slide-3
SLIDE 3

Camera-Robot Configurations

Image from: CHANG, W., WU, C.. Hand-Eye Coordination for Robotic Assembly Tasks. International Journal of Automation and Smart Technology,

slide-4
SLIDE 4

Image-based visual servoing

Current Image Goal Image

slide-5
SLIDE 5

Camera Motion to Image Motion

vx vy vz ωz ωx ωy

Slides adapted from Peter Corke

slide-6
SLIDE 6

The Image Jacobian

ω v ˆ f = f ρ ( ˙ u, ˙ v)T (X, Y, Z)T

✓ ˙ u ˙ v ◆ = ✓ − ˆ f/Z u/Z uv/ ˆ f −( ˆ f + u2/ ˆ f) v − ˆ f/Z v/Z ˆ f + u2/ ˆ f −uv/ ˆ f −u ◆ B B B B B B @ vx vy vz ωx ωy ωz 1 C C C C C C A Slides adapted from Peter Corke

The Image Jacobian

ω v ˆ f = f ρ ( ˙ u, ˙ v)T (X, Y, Z)T

✓ ˙ u ˙ v ◆ = ✓ − ˆ f/Z u/Z uv/ ˆ f −( ˆ f + u2/ ˆ f) v − ˆ f/Z v/Z ˆ f + u2/ ˆ f −uv/ ˆ f −u ◆ B B B B B B @ vx vy vz ωx ωy ωz 1 C C C C C C A Slides adapted from Peter Corke

slide-7
SLIDE 7

Optical flow Patterns

Slides adapted from Peter Corke

slide-8
SLIDE 8

Image-based visual servoing

Getting a camera velocity to minimize the error between the current and goal image

Current Image Goal Image

slide-9
SLIDE 9

Image-based visual servoing

Current Image Goal Image

✓ ˙ u ˙ v ◆ = ✓ − ˆ f/Z u/Z uv/ ˆ f −( ˆ f + u2/ ˆ f) v − ˆ f/Z v/Z ˆ f + u2/ ˆ f −uv/ ˆ f −u ◆ B B B B B B @ vx vy vz ωx ωy ωz 1 C C C C C C A

J(u, v, Z)

Slides adapted from Peter Corke

slide-10
SLIDE 10

Image-based visual servoing

Current Image Goal Image

        ˙ u1 ˙ v1 ˙ u2 ˙ v2 ˙ u3 ˙ v3         =   J(u1, v1, Z1) J(u2, v2, Z2) J(u3, v3, Z3)           vx vy vz ωx ωy ωz        

slide-11
SLIDE 11

Image-based visual servoing

        ˙ u1 ˙ v1 ˙ u2 ˙ v2 ˙ u3 ˙ v3         =   J(u1, v1, Z1) J(u2, v2, Z2) J(u3, v3, Z3)           vx vy vz ωx ωy ωz        

        vx vy vz ωx ωy ωz         =   J(u1, v1, Z1) J(u2, v2, Z2) J(u3, v3, Z3)  

−1

        ˙ u1 ˙ v1 ˙ u2 ˙ v2 ˙ u3 ˙ v3        

slide-12
SLIDE 12

Desired Pixel Velocity

Slides adapted from Peter Corke

slide-13
SLIDE 13

Simulation

Slides adapted from Peter Corke

slide-14
SLIDE 14

Point Correspondences

How to find them? Features, Markers

slide-15
SLIDE 15

What will you take home today?

Visual Servoing Interaction Matrix Control Law Case-Study: Learning-based approach Introduction Optimal Control Principle of Optimality Bellman Equation Deriving LQR

slide-16
SLIDE 16

Training Deep Neural Networks for Visual Servoing

Bateux et al. ICRA 2018

1.

Instead of using features, use the whole image to compare to given goal image

  • a. Challenge: Small convergence region due to non-linear cost function
slide-17
SLIDE 17

Training Deep Neural Networks for Visual Servoing

Bateux et al. ICRA 2018

slide-18
SLIDE 18

Training Deep Neural Networks for Visual Servoing

Bateux et al. ICRA 2018

slide-19
SLIDE 19

What will you take home today?

Visual Servoing Interaction Matrix Control Law Case-Study: Learning-based approach Introduction Optimal Control Principle of Optimality Bellman Equation Deriving LQR

slide-20
SLIDE 20

So far on Control

slide-21
SLIDE 21

Optimal Control and Reinforcement Learning from a unified point of view

Optimal Control Problem

slide-22
SLIDE 22

Trajectory Optimization and Reinforcement Learning

1.

Trajectory Optimization: find a optimal trajectory given non-linear dynamics and cost

2.

Reinforcement Learning: finding an optimal policy under unknown dynamics and given a reward = -cost

slide-23
SLIDE 23

Principle of Optimality – Example: Graph Search problem

slide-24
SLIDE 24

Forward Search

slide-25
SLIDE 25

Forward Search

slide-26
SLIDE 26

Backward Search

slide-27
SLIDE 27

Backward Search

slide-28
SLIDE 28

Principle of Optimality

slide-29
SLIDE 29

Bellman Equation

slide-30
SLIDE 30

Problem setup

System Dynamics Cost function

slide-31
SLIDE 31

Goal

slide-32
SLIDE 32

Formalize Cost-to-Go / Value function

slide-33
SLIDE 33

Optimal Value function = V with lowest cost

slide-34
SLIDE 34

Deriving the Bellman Equation

slide-35
SLIDE 35

Optimal Bellman Equation

Optimal Value function Optimal Policy

slide-36
SLIDE 36

Comparing Optimal Bellman and Value Function

slide-37
SLIDE 37

Infinite time horizon, deterministic system

slide-38
SLIDE 38

Infinite time horizon, deterministic system

slide-39
SLIDE 39

Infinite time horizon, deterministic system

slide-40
SLIDE 40

Finite Horizon, Stochastic system

Stochastic System Dynamics Cost function

slide-41
SLIDE 41

Finite Horizon, Stochastic system

Value function Optimal Value function Optimal Policy

slide-42
SLIDE 42

Finite Horizon, Stochastic system

Bellman Equation Optimal Bellman Equation

slide-43
SLIDE 43

Infinite Horizon, Stochastic system

Combining formulation from infinite horizon - discrete system with stochastic system derivation

slide-44
SLIDE 44

Continuous time systems

Hamilton-Jacobi-Bellman Equation

slide-45
SLIDE 45

How do you solve these equations?

slide-46
SLIDE 46

Linear Dynamical Systems, Quadratic cost – Linear Quadratic Regulator (LQR)

slide-47
SLIDE 47

Linear Dynamical Systems, Quadratic cost – Linear Quadratic Regulator (LQR)