Sensitivity of Joint Estimation in Multi Agent Iterative Learning - - PowerPoint PPT Presentation

sensitivity of joint estimation in multi agent iterative
SMART_READER_LITE
LIVE PREVIEW

Sensitivity of Joint Estimation in Multi Agent Iterative Learning - - PowerPoint PPT Presentation

Sensitivity of Joint Estimation in Multi Agent Iterative Learning Control Angela Schoellig and Raffaello DAndrea Institute for Dynamic Systems and Control ETH Zurich, Switzerland 1 IFAC World Congress 2011, Milano Aug 29, 2011 OUR FOCUS


slide-1
SLIDE 1

1

Sensitivity of Joint Estimation in Multi‐Agent Iterative Learning Control

Angela Schoellig and Raffaello D‘Andrea

Institute for Dynamic Systems and Control ETH Zurich, Switzerland IFAC World Congress 2011, Milano – Aug 29, 2011

slide-2
SLIDE 2

2

OUR FOCUS

  • Group of similar agents
  • Individual agents learn to perform a single‐agent task
  • The task: learn to follow a trajectory
  • Does sharing information speed up simultaneous learning?

Angela Schoellig ‐ ETH Zurich

slide-3
SLIDE 3

3

AGENTS ARE ABLE TO LEARN...

Trajectory tracking with a quadrocopter.

Full-length video. www.tiny.cc/QuadroLearnsTrajectory [Schoellig and D'Andrea, ECC 2009] [Schoellig, Mueller and D'Andrea, submitted to Autonomous Robots]

slide-4
SLIDE 4

4

…when learning the same task?

CAN AGENTS BENEFIT FROM EACH OTHER...

Angela Schoellig ‐ ETH Zurich

slide-5
SLIDE 5

5 Physical model of real‐world system.

Group of similar agents.

Same nominal dynamics

Performing the same task. Repeated and simultaneous operation.

LEARNING OF OPEN‐LOOP CONTROL CORRECTIONS.

PROBLEM STATEMENT

Angela Schoellig ‐ ETH Zurich

Q1: Is an individual agent able to learn faster when performing a task

simultaneously with a group of similar agents?

GOAL OF LEARNING: Follow the desired trajectory.

slide-6
SLIDE 6

6

  • Linearize. Small deviations from nominal trajectory.
  • Discretize. Linear, time‐varying difference equations.

Lifted‐system representation. Static mapping representing one execution. With and

LIFTED‐DOMAIN REPRESENTATION

Angela Schoellig ‐ ETH Zurich

slide-7
SLIDE 7

7

For trial and agent Repetitive disturbance. Unknown. Constant over iterations.

  • Noise. Unknown. Uncorrelated between iterations.

SIMILAR BUT NOT IDENTICAL...

Angela Schoellig ‐ ETH Zurich Agents differ in the unknown part. SIMILARITY ASSUMPTION.

Over iterations our knowledge on and changes…

slide-8
SLIDE 8

8 (1) Estimate the repetitive disturbance by taking into account all past measurements. Obtain . (2) Correct for by updating the input. “Minimize” . For example,

HOW DOES A SINGLE AGENT LEARN?

NEW ITERATION

EXECUTE ESTIMATE CORRECT

Angela Schoellig ‐ ETH Zurich

Can the disturbance estimate be improved by taking into account the measurements of the other agents?

slide-9
SLIDE 9

9

FOCUS: ESTIMATION PROBLEM

Angela Schoellig ‐ ETH Zurich

INDEPENDENT ESTIMATION vs. JOINT ESTIMATION

slide-10
SLIDE 10

10

REDUCE MODEL

Angela Schöllig ‐ ETH Zürich with MEASUREMENT AND PROCESS NOISE with

DYNAMICS

 neglect deterministic part  assume independence of vector entries

slide-11
SLIDE 11

11

Kalman filter for the joint problem.

Estimation objective: System equation: Initial condition:

LEMMA: We obtain covariance matrix in closed form. (Proof by induction)

Special case: independent estimation

JOINT ESTIMATION

Angela Schoellig ‐ ETH Zurich SIMILARITY ASSUMPTION.

slide-12
SLIDE 12

12

JOINT LEARNING BENEFIT METRIC: ratio of state covariances of independent vs. joint estimation

COMPARISON

Angela Schoellig ‐ ETH Zurich

The VARIANCE OF THE STATE ESTIMATE is a measure for the learning performance (=experimental outcome).

If R > 1, joint learning is beneficial.

with

slide-13
SLIDE 13

13

Performance increase due to joint estimation:

THEOREM 1: Pure Process Noise THEOREM 2: Pure Measurement Noise

RESULT

Angela Schoellig ‐ ETH Zurich

limit case for limit case for

[Schoellig, Alonso-Mora and D'Andrea; CDC 2010, accepted AJC]

slide-14
SLIDE 14

14

Under the given assumptions, joint estimation...

  • improves the performance of an individual agent
  • the benefit is only significant if

(1) agents are highly similar AND (2) process noise is negligible AND (3) common disturbance large compared to the measurement noise

SUMMARY

Angela Schoellig ‐ ETH Zurich

Q2: How critical is the underlying similarity assumption?

slide-15
SLIDE 15

15

True values. For

ASSUME THAT DEGREE OF SIMILARITY IS UNKNOWN.

Nominal values (“our best guess“).

SIMILARITY ASSUMPTION

Angela Schoellig ‐ ETH Zurich

Defines degree of similarity.

SOLVE KALMAN FILTER EQUATIONS UNDER NEW ASSUMPTIONS

slide-16
SLIDE 16

16

JOINT ESTIMATION PERFORMANCE IS DEGRADED. LEMMA: Sufficient condition Worst case. Assume agents are identical and they are not, then joint estimation does NOT converge.

SENSITIVITY ANALYSIS − RESULTS

Angela Schoellig ‐ ETH Zurich

Underestimate similarity  Joint estimation remains beneficial.

slide-17
SLIDE 17

17

CONCLUSION

In the proposed framework, where we learn open‐loop input corrections...

TAKE HOME MESSAGE: (1) Joint learning good only if high similarity of unknown disturbance can be guaranteed (2) For joint learning, it‘s always safer to underestimate similarity.

Angela Schoellig ‐ ETH Zurich

Choose independent learning as default since benefit of joint learning is minor for most cases.

slide-18
SLIDE 18

18

Sensitivity of Joint Estimation in Multi‐Agent Iterative Learning Control

Angela Schoellig and Raffaello D‘Andrea

Institute for Dynamic Systems and Control ETH Zurich, Switzerland IFAC World Congress 2011, Milano – Aug 29, 2011