deep robotic learning
play

Deep Robotic Learning Sergey Levine UC Berkeley Google Brain - PowerPoint PPT Presentation

Deep Robotic Learning Sergey Levine UC Berkeley Google Brain robotic state low-level modeling & control observations estimation planning controls control prediction (e.g. vision) pipeline standard classifier features


  1. Deep Robotic Learning Sergey Levine UC Berkeley Google Brain

  2. robotic state low-level modeling & control observations estimation planning controls control prediction (e.g. vision) pipeline

  3. standard classifier features mid-level features computer (e.g. SVM) (e.g. HOG) (e.g. DPM) vision Felzenszwalb ‘08 end-to-end training deep learning robotic state low-level modeling & control controls observations estimation planning control prediction (e.g. vision) pipeline end-to-end training deep state low-level modeling & robotic controls observations estimation planning prediction control (e.g. vision) learning

  4. no direct supervision actions have consequences

  5. 1. Does end-to-end learning produce better sensorimotor skills? 2. Can we apply sensorimotor skill learning to a wide variety of robots & tasks? 3. Can we scale up deep robotic learning and produce skills that generalize ? 4. How can we learn safely and efficiently in safety-critical domains? 5. Can we transfer skills from simulation to the real world , and from one robot to another ?

  6. 1. Does end-to-end learning produce better sensorimotor skills? 2. Can we apply sensorimotor skill learning to a wide variety of robots & tasks? 3. Can we scale up deep robotic learning and produce skills that generalize ? 4. How can we learn safely and efficiently in safety-critical domains? 5. Can we transfer skills from simulation to the real world , and from one robot to another ?

  7. Chelsea Finn

  8. 96.3% success rate end-to-end training 0% success (trained on pose only) rate pose prediction L.*, Finn*, Darrell, Abbeel , ‘16

  9. 1. Does end-to-end learning produce better sensorimotor skills? 2. Can we apply sensorimotor skill learning to a wide variety of robots & tasks? 3. Can we scale up deep robotic learning and produce skills that generalize ? 4. How can we learn safely and efficiently in safety-critical domains? 5. Can we transfer skills from simulation to the real world , and from one robot to another ?

  10. Deep Robotic Learning Applications manipulation dexterous hands soft hands with N. Wagener, P. Abbeel with V. Kumar, A. Gupta, E. Todorov with C. Eppner, A. Gupta, P. Abbeel locomotion aerial vehicles tensegrity robot with X. Geng, M. Zhang, J. Bruce, K. Caluwaerts, M. Vespignani, V. SunSpiral, P. Abbeel with G. Kahn, T. Zhang, P. Abbeel with V. Koltun

  11. 1. Does end-to-end learning produce better sensorimotor skills? 2. Can we apply sensorimotor skill learning to a wide variety of robots & tasks? 3. Can we scale up deep robotic learning and produce skills that generalize ? 4. How can we learn safely and efficiently in safety-critical domains? 5. Can we transfer skills from simulation to the real world , and from one robot to another ?

  12. ingredients for success in learning: supervised learning: learning robotic skills: computation computation ~ data algorithms algorithms ? data

  13. Grasping with Learned Hand-Eye Coordination monocular • monocular camera (no depth) RGB camera • no camera calibration either 7 DoF arm • 2-5 Hz update • continuous arm control 2-finger gripper • servo the gripper to target • fix mistakes object • no prior knowledge bin Alex Peter Pastor Krizhevsky Deirdre Quillen L., Pastor, Krizhevsky, Quillen ‘16

  14. Grasping Experiments

  15. Policy Learning with Multiple Robots Rollout execution Local policy optimization Global policy optimization Mrinal Ali Yahya Kalakrishnan Yevgen Chebotar Adrian Li

  16. Yahya, Li, Kalakrishnan, Chebotar , L., ‘16

  17. Policy Learning with Multiple Robots: Deep RL with NAF Ethan Holly Tim Lillicrap Shane Gu Gu*, Holly*, Lillicrap , L., ‘16

  18. Learning a Predictive Model of Natural Images original video Chelsea Finn predictions

  19. 1. Does end-to-end learning produce better sensorimotor skills? 2. Can we apply sensorimotor skill learning to a wide variety of robots & tasks? 3. Can we scale up deep robotic learning and produce skills that generalize ? 4. How can we learn safely and efficiently in safety-critical domains? 5. Can we transfer skills from simulation to the real world , and from one robot to another ?

  20. Safe Uncertainty-Aware Learning unknown environment Key idea: To learn about collisions, must experience collisions (but safely!) 1. Learn a collision prediction model raw image command velocities neural network ensemble 2. Speed-dependent, uncertainty-aware collision cost Greg Kahn 3. Iteratively train with on-policy samples Kahn, Pong, Abbeel , L. ‘16

  21. Safe Uncertainty-Aware Learning Kahn, Pong, Abbeel , L. ‘16

  22. 1. Does end-to-end learning produce better sensorimotor skills? 2. Can we apply sensorimotor skill learning to a wide variety of robots & tasks? 3. Can we scale up deep robotic learning and produce skills that generalize ? 4. How can we learn safely and efficiently in safety-critical domains? 5. Can we transfer skills from simulation to the real world , and from one robot to another ?

  23. Training in Simulation: CAD2RL Fereshteh Sadeghi Sadeghi , L. ‘16

  24. Training in Simulation: CAD2RL Sadeghi , L. ‘16

  25. Training in Simulation: CAD2RL Sadeghi , L. ‘16

  26. Sadeghi , L. ‘16

  27. Learning with Transfer in Mind: Ensemble Policy Optimization (EPOpt) training on single torso mass training on model ensemble train test unmodeled effects ensemble adaptation adapt Aravind Rajeswaran

  28. 1. Does end-to-end learning produce better sensorimotor skills? 2. Can we apply sensorimotor skill learning to a wide variety of robots & tasks? 3. Can we scale up deep robotic learning and produce skills that generalize ? 4. How can we learn safely and efficiently in safety-critical domains? 5. Can we transfer skills from simulation to the real world , and from one robot to another ? 6. How can we get sufficient supervision to learn in unstructured real-world environments?

  29. Learning what Success Means can we learn the goal with visual features? Finn, Abbeel , L. ‘16

  30. Learning what Success Means Sermanet , Xu, L. ‘16

  31. ingredients for success in learning: supervised learning: learning robotic skills: computation computation ~ data algorithms algorithms ? data

  32. Fereshteh Sadeghi Aravind Rajeswaran Chelsea Finn Greg Kahn Announcement: New Conference Conference on Robotic Learning (CoRL) www.robot-learning.org Goal: bring together robotics & machine learning in a focused conference format Alex Conference: November 2017 Peter Pastor Krizhevsky Deirdre Quillen Trevor Darrell Pieter Abbeel Papers deadline: late June 2017 Steering committee: Ken Goldberg (UC Berkeley), Sergey Levine (UC Berkeley), Vincent Vanhoucke (Google), Abhinav Gupta (CMU), Stefan Schaal (USC, MPI), Michael I. Jordan (UC Berkeley), Raia Hadsell (DeepMind), Dieter Fox (UW), Joelle Pineau (McGill), J. Andrew Bagnell (CMU), Aude Billard (EPFL), Stefanie Tellex (Brown), Minoru Asada (Osaka), Wolfram Burgard (Freiburg), Pieter Abbeel (UC Berkeley) Mrinal Shane Gu Ethan Holly Tim Lillicrap Kalakrishnan Yevgen Chebotar Ali Yahya Adrian Li

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend