visual grounding of learned physical models
play

Visual Grounding of Learned Physical Models ICML 2020 Yunzhu Li - PowerPoint PPT Presentation

Visual Grounding of Learned Physical Models ICML 2020 Yunzhu Li Toru Lin* Kexin Yi* Daniel M. Bear Daniel L.K. Yamins Jiajun Wu Joshua B. Antonio Torralba Tenenbaum http://visual-physics-grounding.csail.mit.edu/ (* indicates equal


  1. Visual Grounding of Learned Physical Models ICML 2020 Yunzhu Li Toru Lin* Kexin Yi* Daniel M. Bear Daniel L.K. Yamins Jiajun Wu Joshua B. Antonio Torralba Tenenbaum http://visual-physics-grounding.csail.mit.edu/ (* indicates equal contribution)

  2. Intuitive Physics (1) Distinguish between different instances (2) Recognize objects’ physical properties (3) Predict future movements (Wu et al., Learning to See Physics via Visual De-animation)

  3. Intuitive Physics (1) Distinguish between different instances (2) Recognize objects’ physical properties (3) Predict future movements (Wu et al., Learning to See Physics via Visual De-animation)

  4. Larger stiffness Smaller stiffness For example Different physical parameters lead to different motions. Estimating physical parameter Larger gravity Smaller gravity by comparing mental simulation with observation

  5. Physical reasoning of deformable objects is challenging.

  6. Physical reasoning of deformable objects is challenging. Particle-based Representation General & Flexible

  7. Physical reasoning of deformable objects is challenging. Particle-based Representation General & Flexible We propose a model that jointly (1) Estimates the physical properties (2) Refines the particle locations using (1) a learned visual prior (2) a learned dynamics prior

  8. Visually Grounded Physics Learner (VGPL)

  9. Visually Grounded Physics Learner (VGPL)

  10. Visually Grounded Physics Learner (VGPL) Visual Grounding

  11. Visually Grounded Physics Learner (VGPL) Visual Grounding

  12. We evaluate our model in environments involving interactions between rigid objects, elastic materials, and fluids.

  13. We evaluate our model in environments involving interactions between rigid objects, elastic materials, and fluids. Within a few observation steps, our model is able to (1) refine the state estimation and reason about the physical properties (2) make predictions into the future.

  14. Related Work Learning-based particle dynamics Mrowca, Zhuang, Wang, Haber, Fei-Fei, Battaglia, Pascanu, Lai, Rezende, Tenenbaum, Yamins. NeurIPS’18 Li, Wu, Tedrake, Tenenbaum, Torralba. ICLR’19 Kavukcuoglu. NeurIPS’16 Sanchez-Gonzalez, Godwin, Pfaff, Ying, Leskovec, Ummenhofer, Prantl, Thuerey, Koltun. ICLR’20 Battaglia. ICML’20

  15. Related Work Questions remains: (1) How well they handle visual inputs? Learning-based particle dynamics (2) How to adapt to scenarios of unknown physical parameters? Mrowca, Zhuang, Wang, Haber, Fei-Fei, Battaglia, Pascanu, Lai, Rezende, Tenenbaum, Yamins. NeurIPS’18 Li, Wu, Tedrake, Tenenbaum, Torralba. ICLR’19 Kavukcuoglu. NeurIPS’16 Sanchez-Gonzalez, Godwin, Pfaff, Ying, Leskovec, Ummenhofer, Prantl, Thuerey, Koltun. ICLR’20 Battaglia. ICML’20

  16. Related Work Differentiating through physics-based simulators Hu, Liu, Spielberg, Tenenbaum, Schenck, Fox. CoRL’18 Freeman, Wu, Rus, Matusik. ICRA’19 Liang, Lin, Koltun. NeurIPS’19 Belbute-Peres, Smith, Allen, Tenenbaum, Kolter. NeurIPS’18 Degrave, Hermans, Dambre, Wyffels. Frontiers in Neurorobotics 2019

  17. Related Work Questions remains: (1) Make strong assumptions on the Differentiating through physics-based simulators structure of the system (2) Usually time-consuming (2) Prone to local optimum (3) Lacking ways to handle visual inputs Hu, Liu, Spielberg, Tenenbaum, Schenck, Fox. CoRL’18 Freeman, Wu, Rus, Matusik. ICRA’19 Liang, Lin, Koltun. NeurIPS’19 Belbute-Peres, Smith, Allen, Tenenbaum, Kolter. NeurIPS’18 Degrave, Hermans, Dambre, Wyffels. Frontiers in Neurorobotics 2019

  18. Our Work We proposed Visually Grounded Physics Learner (VGPL) to (1) bridge the perception gap, (2) enable physical reasoning from visual perception, and (3) perform dynamics-guided inference to directly predict the optimization results, which allows quick adaptation to environments with unknown physical properties.

  19. Problem Formulation Consider a system that contains objects and particles.

  20. Problem Formulation Consider a system that contains objects and particles. : Visual observ.

  21. Problem Formulation Consider a system that contains objects and particles. Visual prior : Visual observ. : Particle position : Instance grouping

  22. Problem Formulation Consider a system that contains objects and particles. Visual prior Dynamics prior : Visual observ. : Particle position : Instance grouping

  23. Problem Formulation Consider a system that contains objects and particles. Visual prior Dynamics prior : Visual observ. : Particle position : Instance grouping : Rigidness of each instance

  24. Problem Formulation Consider a system that contains objects and particles. Visual prior Dynamics prior : Visual observ. : Particle position : Instance grouping : Rigidness of each instance : Physical parameters

  25. Problem Formulation Consider a system that contains objects and particles. Visual prior Dynamics prior Inference module : Visual observ. : Particle position : Instance grouping : Rigidness of each instance : Physical parameters

  26. Problem Formulation Consider a system that contains objects and particles. Visual prior Dynamics prior Inference module : Visual observ. : Particle position : Instance grouping : Rigidness of each instance : Physical parameters : Position refinement

  27. Problem Formulation Consider a system that contains objects and particles. Visual prior Dynamics prior Inference module : Visual observ. : Particle position : Instance grouping : Rigidness of each instance : Physical parameters Objective function : Position refinement

  28. Visual Prior Visual observations :

  29. Visual Prior Visual observations : Particle locations : Instance grouping :

  30. Visual Prior Visual observations : Particle locations : Instance grouping : Objective function

  31. Results of the Visual Prior Visual Inputs Prediction Visual Inputs Prediction

  32. Dynamics Prior : Particle position : Instance grouping

  33. Dynamics Prior : Particle position : Instance grouping : Rigidness of each instance : Physical parameters

  34. Dynamics Prior : Particle position : Instance grouping : Rigidness of each instance : Physical parameters Li, Wu, Tedrake, Tenenbaum, Torralba, “Learning Particle Dynamics for Manipulating Rigid Bodies, Deformable Objects, and Fluids,” ICLR’19

  35. Results of the Dynamics Prior

  36. Dynamics-Guided Inference

  37. Dynamics-Guided Inference : Rigidness of each instance : Physical parameters

  38. Dynamics-Guided Inference : Rigidness of each instance : Physical parameters : Particle position : Instance grouping

  39. Dynamics-Guided Inference : Rigidness of each instance : Physical parameters : Particle position : Instance grouping : Position refinement

  40. Results We will mainly investigate how accurate the following estimations are and whether they help with future prediction: (1) : Rigidness estimation (2) : Parameter estimation (3) : : Position refinement

  41. Qualitative results on Rigidness Estimation

  42. Quantitative results on Rigidness Estimation Mean accuracy Mean accuracy

  43. Qualitative results on Parameter Estimation

  44. Quantitative results on Parameter Estimation

  45. Qualitative results on Position Refinement

  46. Quantitative results on Position Refinement

  47. Quantitative results on Future Prediction

  48. In summary We proposed Visually Grounded Physics Learner (VGPL) to (1) simultaneously reason about physics and make future predictions based on visual and dynamics priors.

  49. In summary We proposed Visually Grounded Physics Learner (VGPL) to (1) simultaneously reason about physics and make future predictions based on visual and dynamics priors. (2) We employ a particle-based representation to handle rigid bodies, deformable objects, and fluids.

  50. In summary We proposed Visually Grounded Physics Learner (VGPL) to (1) simultaneously reason about physics and make future predictions based on visual and dynamics priors. (2) We employ a particle-based representation to handle rigid bodies, deformable objects, and fluids. (3) Experiments show that our model can infer the physical properties within a few observations, which allows the model to quickly adapt to unseen scenarios and make accurate predictions into the future.

  51. Thank you for watching!

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend