improving tracking performance by learning from past data
play

Improving tracking performance by learning from past data Angela P. - PowerPoint PPT Presentation

Improving tracking performance by learning from past data Angela P. Schoellig Doctoral examina on July 30, 2012 Advisor: Prof. Raffaello DAndrea // Co advisor: Prof. Andrew Alleyne 1 Improving tracking performance by learning from


  1. Improving tracking performance by learning from past data Angela P. Schoellig Doctoral examina � on − July 30, 2012 Advisor: Prof. Raffaello D’Andrea // Co ‐ advisor: Prof. Andrew Alleyne 1

  2. Improving tracking performance by learning from past data Angela P. Schoellig Doctoral examina � on − July 30, 2012 Advisor: Prof. Raffaello D’Andrea // Co ‐ advisor: Prof. Andrew Alleyne 2

  3. MOTIVATION HUMANS learn from experience. We learn from mistakes and get better through practice. We constantly adapt to changing environments. 3

  4. MOTIVATION AUTOMATED SYSTEMS typically make the same mistakes over and over again when performing a task repeatedly. Why? Robots of a car assembly line. 4

  5. MOTIVATION AUTOMATED SYSTEMS are typically operated using feedback control: Disturbance Input Output CONTROLLER PLANT Performance limitations: • Causality of disturbance correction: “first detect error, then react”. • Model ‐ based controller design; model ≠ real system. 5

  6. GOAL Improve the performance over causal, feedback control by learning from previous experiments. Disturbance Input Output SYSTEM LEARNING 6

  7. SCOPE OF WORK Output Input LEARNING Learning task: Following a predefined trajectory. Approach: • Model ‐ based learning based on a priori knowledge of the system dynamics. • Adaptation of the input. Potential: Acausal action, anticipating repetitive disturbances. 7

  8. OVERVIEW I. Introduction a. Testbed: The Flying Machine Arena b. Motivation for learning II. Project A. Iterative learning for precise trajectory following: single ‐ agent and multi ‐ agent results. III. Project B. Learning of feed ‐ forward parameters for rhythmic flight performances IV. Summary 8

  9. TESTBED, see www.flyingmachinearena.org 9

  10. THE TEAM Mark Müller Markus Hehn Federico Augugliaro Sergei Lupashin 10

  11. THE FLYING MACHINE ARENA Vehicle position and attitude Control Algorithms wireless Collective thrust and turn rates (wireless) 11

  12. OPERATION Input Output Trajectory ‐ following Desired position controller (TFC) Collective thrust and turn rates Measured position and attitude 12

  13. MOTIVATION: PROJECT A Desired motion. 13

  14. MOTIVATION: PROJECT A Performance with trajectory ‐ following controller. Different trials Large repetitive error 14

  15. OVERVIEW I. Introduction II. Project A. Iterative learning for precise trajectory following a. Learning approach b. Results III. Project B. Learning of feed ‐ forward parameters for rhythmic flight performances IV. Summary 15

  16. A | PUBLICATIONS Peer ‐ reviewed publications Schoellig, A. P. and R. D’Andrea (2009): “Optimization ‐ based iterative learning control for trajectory tracking.” In Proceedings of the European Control Conference (ECC) . Schoellig, A. P., F. L. Mueller, and R. D’Andrea (2012): “Optimization ‐ based iterative learning for precise quadrocopter trajectory tracking.” Auton ‐ omous Robots. Mueller, F.L., A. P. Schoellig, and R. D’Andrea (2012): “Iterative learning of feed ‐ forward corrections for high ‐ performance tracking.” To appear in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) . Joint work with Fabian L. Mueller (Master student). 16

  17. A | LEARNING APPROACH Features: Learning through a repeated operation, updating full input trajectory after each trial. Input trajectory Output trajectory SYSTEM DISTURBANCE ESTIMATION INPUT UPDATE Updated input Estimated disturbance LEARNING 17

  18. A | LEARNING APPROACH PREREQUISITES Dynamics model of system • (i) in analytical form or (ii) in form of a numerical dynamics simulation Desired output trajectory and corresponding • nominal input trajectory  must satisfy the model equations. RESULT Learned input • Estimated disturbance vector • 18

  19. A | LIFTED ‐ DOMAIN REPRESENTATION Dynamics model of the physical system: Consider small deviations from nominal trajectory. Linearize and discretize. Linear, time ‐ varying difference equation. Static mapping. Representing one trial. With , and . 19

  20. A | ITERATION ‐ DOMAIN MODEL For each trial Recurring disturbance . Unknown. Only small changes between iterations: trial ‐ uncorrelated, zero ‐ mean Gaussian noise Noise . Unknown. Changing from iteration to iteration. From trial to trial our knowledge about improves. 20

  21. A | STEP 1: ESTIMATION UPDATE OF DISTURBANCE ESTIMATE EXECUTE via Kalman filter in the iteration domain: estimates the repetitve disturbance by taking into account all past measurements. ESTIMATE Prediction step: UPDATE Measurement update step: Obtain . 21

  22. A | STEP 2: UPDATE INPUT UPDATE via convex optimization: EXECUTE minimizes the tracking error in the next trial: ESTIMATE UPDATE subject to Obtain . 22

  23. A | TWO EXPERIMENTAL SCENARIOS SCENARIO 1 SCENARIO 2 • No feedback from motion capture • Camera information is used. cameras during task execution Collective thrust Position, Position, and turn rates attitude Position attitude TFC • Analytical model • Model via numerical simulation • 2D quadrocopter model • 3D quadrocopter model • Constraints on single motor thrusts and turn rates. 23

  24. A | SCENARIO 1: state trajectories S ‐ shaped trajectory. 24

  25. A | SCENARIO 1: input trajectories S ‐ shaped trajectory. 25

  26. A | SCENARIO 1: state trajectories S ‐ shaped trajectory. 26

  27. A | TWO EXPERIMENTAL SCENARIOS SCENARIO 1 SCENARIO 2 • No feedback from motion capture • Camera information is used. cameras during task execution Collective thrust Position, Position, and turn rates attitude Position attitude TFC • Analytical model • Model via numerical simulation • 2D quadrocopter model • 3D quadrocopter model • Constraints on single motor thrusts and turn rates. 27

  28. A | SCENARIO 2: state trajectories S ‐ shaped trajectory. 28

  29. A | SCENARIO 2: state trajectories S ‐ shaped trajectory. 29

  30. A | SCENARIO 2: state trajectories S ‐ shaped trajectory. 30

  31. A | SCENARIO 2: state trajectories S ‐ shaped trajectory. 31

  32. A | SCENARIO 2: error convergence 32

  33. A | SUMMARY • Prerequisites: approximate model of system dynamics. • Efficient learning algorithm: convergence in around 5 ‐ 10 iterations. Acausal compensation: outperforms pure feedback control. • Scenario 2: without learning with learning Powerful combination Learning applied to feedback ‐ control systems: compensation for repetitive and non ‐ repetitive disturbances. 33

  34. VIDEO: http://tiny.cc/SlalomLearning 34

  35. OVERVIEW I. Introduction II. Project A. Iterative learning for precise trajectory following III. Project B. Learning of feed ‐ forward parameters for rhythmic flight performances a. Learning approach b. Results I. Summary 35

  36. B | PUBLICATIONS Peer ‐ reviewed publications Schoellig, A. P., F. Augugliaro, and R. D’Andrea (2009): “Synchronizing the motion of a quadrocopter to music.” In Proceedings of IEEE International Conference on Robotics and Automation (ICRA). Schoellig, A. P., F. Augugliaro, and R. D’Andrea (2010): “A platform for dance performances with multiple quadrocopters.” In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS)–Workshop on Robots and Musical Expressions. Schoellig, A. P., M. Hehn, S. Lupashin, and R. D’Andrea (2011): “Feasibility of motion primitives for choreographed quadrocopter flight.” In Proceedings of the American Control Conference (ACC) . Schoellig, A. P., C. Wiltsche, and R. D’Andrea (2012): “Feed ‐ forward parameter identification for precise periodic quadrocopter motions.” In Proceedings of the American Control Conference (ACC). Joint work with Federico Augugliaro (Bachelor/Master student) and Clemens Wiltsche (semester project). 36

  37. VIDEO: http://tiny.cc/DanceWith3 37

  38. B | LEARNING APPROACH Task: Precise tracking of periodic motions. Features: • Learning through a dedicated identification routine performed prior to flight performance. • Adaptation of only a few input parameters. Position, Position attitude TFC 38

  39. B | LEARNING APPROACH Amplitude and phase error PURE FEEDBACK WITH LEARNED CORRECTION FACTORS For each directional motion component and frequency, we learn: (1) amplitude correction factor, (2) additive phase correction. 39

  40. VIDEO: http://tiny.cc/Armageddon Angela Schoellig ‐ ETH Zurich 40

  41. OVERVIEW I. Introduction II. Project A. Iterative learning for precise trajectory following III. Project B. Learning of feed ‐ forward parameters for rhythmic flight performances IV. Summary 41

  42. SUMMARY Repetitive error components can be effectively compensated for by learning from past data. Result is an improved tracking performance . Disturbance Output Input SYSTEM LEARNING 42

  43. RESEARCH SUPPORT STAFF Igor Thommen Carolina Flores Hans Ulrich Honegger Marc Corzillius 43

  44. IT FOLLOWS... Live demonstration in the Flying Machine Arena 44

  45. Improving tracking performance by learning from past data Angela P. Schoellig Doctoral examina � on − July 30, 2012 Advisor: Prof. Raffaello D’Andrea // Co ‐ advisor: Prof. Andrew Alleyne 45

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend