Visual Grounding of Learned Physical Models ICML 2020 Yunzhu Li - - PowerPoint PPT Presentation

visual grounding of learned physical models
SMART_READER_LITE
LIVE PREVIEW

Visual Grounding of Learned Physical Models ICML 2020 Yunzhu Li - - PowerPoint PPT Presentation

Visual Grounding of Learned Physical Models ICML 2020 Yunzhu Li Toru Lin* Kexin Yi* Daniel M. Bear Daniel L.K. Yamins Jiajun Wu Joshua B. Antonio Torralba Tenenbaum http://visual-physics-grounding.csail.mit.edu/ (* indicates equal


slide-1
SLIDE 1

Visual Grounding of Learned Physical Models

http://visual-physics-grounding.csail.mit.edu/ (* indicates equal contribution)

Yunzhu Li Toru Lin* Kexin Yi* Daniel M. Bear Daniel L.K. Yamins Jiajun Wu Joshua B. Tenenbaum Antonio Torralba

ICML 2020

slide-2
SLIDE 2

Intuitive Physics

(1) Distinguish between different instances (2) Recognize objects’ physical properties (3) Predict future movements

(Wu et al., Learning to See Physics via Visual De-animation)

slide-3
SLIDE 3

Intuitive Physics

(1) Distinguish between different instances (2) Recognize objects’ physical properties (3) Predict future movements

(Wu et al., Learning to See Physics via Visual De-animation)

slide-4
SLIDE 4

For example

Different physical parameters lead to different motions. Estimating physical parameter by comparing mental simulation with observation

Larger stiffness Smaller stiffness Larger gravity Smaller gravity

slide-5
SLIDE 5

Physical reasoning of deformable objects is challenging.

slide-6
SLIDE 6

Physical reasoning of deformable objects is challenging. Particle-based Representation General & Flexible

slide-7
SLIDE 7

Physical reasoning of deformable objects is challenging. Particle-based Representation General & Flexible We propose a model that jointly (1) Estimates the physical properties (2) Refines the particle locations using (1) a learned visual prior (2) a learned dynamics prior

slide-8
SLIDE 8

Visually Grounded Physics Learner (VGPL)

slide-9
SLIDE 9

Visually Grounded Physics Learner (VGPL)

slide-10
SLIDE 10

Visually Grounded Physics Learner (VGPL)

Visual Grounding

slide-11
SLIDE 11

Visually Grounded Physics Learner (VGPL)

Visual Grounding

slide-12
SLIDE 12

We evaluate our model in environments involving interactions between rigid objects, elastic materials, and fluids.

slide-13
SLIDE 13

We evaluate our model in environments involving interactions between rigid objects, elastic materials, and fluids. Within a few observation steps, our model is able to (1) refine the state estimation and reason about the physical properties (2) make predictions into the future.

slide-14
SLIDE 14

Related Work

Li, Wu, Tedrake, Tenenbaum, Torralba. ICLR’19 Mrowca, Zhuang, Wang, Haber, Fei-Fei, Tenenbaum, Yamins. NeurIPS’18

Learning-based particle dynamics

Sanchez-Gonzalez, Godwin, Pfaff, Ying, Leskovec,

  • Battaglia. ICML’20

Ummenhofer, Prantl, Thuerey, Koltun. ICLR’20 Battaglia, Pascanu, Lai, Rezende,

  • Kavukcuoglu. NeurIPS’16
slide-15
SLIDE 15

Related Work

Li, Wu, Tedrake, Tenenbaum, Torralba. ICLR’19 Mrowca, Zhuang, Wang, Haber, Fei-Fei, Tenenbaum, Yamins. NeurIPS’18

Learning-based particle dynamics

Sanchez-Gonzalez, Godwin, Pfaff, Ying, Leskovec,

  • Battaglia. ICML’20

Ummenhofer, Prantl, Thuerey, Koltun. ICLR’20 Battaglia, Pascanu, Lai, Rezende,

  • Kavukcuoglu. NeurIPS’16

Questions remains: (1) How well they handle visual inputs? (2) How to adapt to scenarios of unknown physical parameters?

slide-16
SLIDE 16

Related Work

Hu, Liu, Spielberg, Tenenbaum, Freeman, Wu, Rus, Matusik. ICRA’19

Differentiating through physics-based simulators

Schenck, Fox. CoRL’18 Degrave, Hermans, Dambre, Wyffels. Frontiers in Neurorobotics 2019 Belbute-Peres, Smith, Allen, Tenenbaum, Kolter. NeurIPS’18 Liang, Lin, Koltun. NeurIPS’19

slide-17
SLIDE 17

Related Work

Hu, Liu, Spielberg, Tenenbaum, Freeman, Wu, Rus, Matusik. ICRA’19

Differentiating through physics-based simulators

Schenck, Fox. CoRL’18 Degrave, Hermans, Dambre, Wyffels. Frontiers in Neurorobotics 2019 Belbute-Peres, Smith, Allen, Tenenbaum, Kolter. NeurIPS’18 Liang, Lin, Koltun. NeurIPS’19

Questions remains: (1) Make strong assumptions on the structure of the system (2) Usually time-consuming (2) Prone to local optimum (3) Lacking ways to handle visual inputs

slide-18
SLIDE 18

Our Work

We proposed Visually Grounded Physics Learner (VGPL) to (1) bridge the perception gap, (2) enable physical reasoning from visual perception, and (3) perform dynamics-guided inference to directly predict the optimization results, which allows quick adaptation to environments with unknown physical properties.

slide-19
SLIDE 19

Problem Formulation

Consider a system that contains objects and particles.

slide-20
SLIDE 20

Problem Formulation

Consider a system that contains objects and particles. : Visual observ.

slide-21
SLIDE 21

Problem Formulation

Consider a system that contains objects and particles. Visual prior : Particle position : Instance grouping : Visual observ.

slide-22
SLIDE 22

Problem Formulation

Consider a system that contains objects and particles. Visual prior Dynamics prior : Particle position : Instance grouping : Visual observ.

slide-23
SLIDE 23

Problem Formulation

Consider a system that contains objects and particles. Visual prior Dynamics prior : Particle position : Instance grouping : Visual observ. : Rigidness of each instance

slide-24
SLIDE 24

Problem Formulation

Consider a system that contains objects and particles. Visual prior Dynamics prior : Particle position : Instance grouping : Visual observ. : Rigidness of each instance : Physical parameters

slide-25
SLIDE 25

Problem Formulation

Consider a system that contains objects and particles. Visual prior Dynamics prior Inference module : Particle position : Instance grouping : Rigidness of each instance : Physical parameters : Visual observ.

slide-26
SLIDE 26

Problem Formulation

Consider a system that contains objects and particles. Visual prior Dynamics prior Inference module : Particle position : Instance grouping : Rigidness of each instance : Physical parameters : Position refinement : Visual observ.

slide-27
SLIDE 27

Problem Formulation

Consider a system that contains objects and particles. Visual prior Dynamics prior Inference module Objective function : Particle position : Instance grouping : Rigidness of each instance : Physical parameters : Position refinement : Visual observ.

slide-28
SLIDE 28

Visual Prior

Visual observations :

slide-29
SLIDE 29

Visual Prior

Particle locations : Instance grouping : Visual observations :

slide-30
SLIDE 30

Visual Prior

Particle locations : Instance grouping : Objective function Visual observations :

slide-31
SLIDE 31

Results of the Visual Prior

Visual Inputs Visual Inputs Prediction Prediction

slide-32
SLIDE 32

Dynamics Prior

: Particle position : Instance grouping

slide-33
SLIDE 33

Dynamics Prior

: Particle position : Instance grouping : Rigidness of each instance : Physical parameters

slide-34
SLIDE 34

Dynamics Prior

: Particle position : Instance grouping : Rigidness of each instance : Physical parameters

Li, Wu, Tedrake, Tenenbaum, Torralba, “Learning Particle Dynamics for Manipulating Rigid Bodies, Deformable Objects, and Fluids,” ICLR’19

slide-35
SLIDE 35

Results of the Dynamics Prior

slide-36
SLIDE 36

Dynamics-Guided Inference

slide-37
SLIDE 37

Dynamics-Guided Inference

: Rigidness of each instance : Physical parameters

slide-38
SLIDE 38

Dynamics-Guided Inference

: Particle position : Instance grouping : Rigidness of each instance : Physical parameters

slide-39
SLIDE 39

Dynamics-Guided Inference

: Particle position : Instance grouping : Rigidness of each instance : Physical parameters : Position refinement

slide-40
SLIDE 40

Results

We will mainly investigate how accurate the following estimations are and whether they help with future prediction: (1) : Rigidness estimation (2) : Parameter estimation (3) : : Position refinement

slide-41
SLIDE 41

Qualitative results on Rigidness Estimation

slide-42
SLIDE 42

Quantitative results on Rigidness Estimation

Mean accuracy Mean accuracy

slide-43
SLIDE 43

Qualitative results on Parameter Estimation

slide-44
SLIDE 44

Quantitative results on Parameter Estimation

slide-45
SLIDE 45

Qualitative results on Position Refinement

slide-46
SLIDE 46

Quantitative results on Position Refinement

slide-47
SLIDE 47

Quantitative results on Future Prediction

slide-48
SLIDE 48

In summary

We proposed Visually Grounded Physics Learner (VGPL) to (1) simultaneously reason about physics and make future predictions based on visual and dynamics priors.

slide-49
SLIDE 49

In summary

We proposed Visually Grounded Physics Learner (VGPL) to (1) simultaneously reason about physics and make future predictions based on visual and dynamics priors. (2) We employ a particle-based representation to handle rigid bodies, deformable objects, and fluids.

slide-50
SLIDE 50

In summary

We proposed Visually Grounded Physics Learner (VGPL) to (1) simultaneously reason about physics and make future predictions based on visual and dynamics priors. (2) We employ a particle-based representation to handle rigid bodies, deformable objects, and fluids. (3) Experiments show that our model can infer the physical properties within a few observations, which allows the model to quickly adapt to unseen scenarios and make accurate predictions into the future.

slide-51
SLIDE 51

Thank you for watching!