bayesian causal induction
play

Bayesian Causal Induction Pedro A. Ortega Sensorimotor Learning and - PowerPoint PPT Presentation

Bayesian Causal Induction Pedro A. Ortega Sensorimotor Learning and Decision-Making Group MPI for Biological Cybernetics/Intelligent Systems 17th December 2011 Introduction Causal Induction (AKA Causal Discovery): One of the oldest


  1. Bayesian Causal Induction Pedro A. Ortega Sensorimotor Learning and Decision-Making Group MPI for Biological Cybernetics/Intelligent Systems 17th December 2011

  2. Introduction Causal Induction (AKA Causal Discovery): ◮ One of the oldest philosophical problems: ◮ Aristotle, Kant, Hume, . . . ◮ The generalization from particular causal instances to abstract causal laws.

  3. Introduction Causal Induction (AKA Causal Discovery): ◮ One of the oldest philosophical problems: ◮ Aristotle, Kant, Hume, . . . ◮ The generalization from particular causal instances to abstract causal laws. ◮ Example: ◮ ‘I had a bad fall on wet floor.’ ◮ ‘Therefore, it is dangerous to ride a bike on ice.’ ◮ (‘Because I learned that a slippery floor can cause a fall’)

  4. Introduction Causal Induction (AKA Causal Discovery): ◮ One of the oldest philosophical problems: ◮ Aristotle, Kant, Hume, . . . ◮ The generalization from particular causal instances to abstract causal laws. ◮ Example: ◮ ‘I had a bad fall on wet floor.’ ◮ ‘Therefore, it is dangerous to ride a bike on ice.’ ◮ (‘Because I learned that a slippery floor can cause a fall’) ◮ Two important aspects: ◮ Infer causal link from experience. ◮ Extrapolate to future experience.

  5. Introduction Causal Induction (AKA Causal Discovery): ◮ One of the oldest philosophical problems: ◮ Aristotle, Kant, Hume, . . . ◮ The generalization from particular causal instances to abstract causal laws. ◮ Example: ◮ ‘I had a bad fall on wet floor.’ ◮ ‘Therefore, it is dangerous to ride a bike on ice.’ ◮ (‘Because I learned that a slippery floor can cause a fall’) ◮ Two important aspects: ◮ Infer causal link from experience. ◮ Extrapolate to future experience. ◮ We all do this in our everyday lives— but how?

  6. Causal Graphical Model � �� � � � � ◮ A pair of (binary) random variables X and Y ◮ Two candidate causal hypotheses { h , ¬ h } (having identical joint distributions)

  7. Causal Graphical Model � �� � � � � ◮ A pair of (binary) random variables X and Y ◮ Two candidate causal hypotheses { h , ¬ h } (having identical joint distributions) ◮ How do we express the problem of causal induction using the language of graphical models alone ?

  8. Causal Graphical Model � � � ◮ A pair of (binary) random variables X and Y ◮ Two candidate causal hypotheses { h , ¬ h } (having identical joint distributions) ◮ How do we express the problem of causal induction using the language of graphical models alone ?

  9. Causal Graphical Model � � � � � ◮ A pair of (binary) random variables X and Y ◮ Two candidate causal hypotheses { h , ¬ h } (having identical joint distributions) ◮ How do we express the problem of causal induction using the language of graphical models alone ? ◮ Do we have to introduce a meta-level for H ?

  10. Probability Trees H 1 1 2 2 � �� X Y � �� � �� 1 1 1 1 2 2 2 2 Y Y X X � �� � �� � �� � �� 3 3 1 1 3 1 1 3 4 4 4 4 4 4 4 4 ◮ Node: mechanism, history dependent ◮ e.g. P ( y | h , ¬ x ) = 1 4 and P ( ¬ y | h , ¬ x ) = 3 4 ◮ Path: causal realization of mechanisms ◮ Tree: causal realizations, possibly heterogeneous ◮ All random variables are first class citizens!

  11. Inferring the Causal Direction ◮ We observe X = x , then we observe Y = y . ◮ What is the probability of H = h ? ◮ Calculate posterior probability: P ( y | h , x ) P ( x | h ) P ( h ) P ( h | x , y ) = P ( y | h , x ) P ( x | h ) P ( h ) + P ( x |¬ h , y ) P ( y |¬ h ) P ( ¬ h ) 4 · 1 3 2 · 1 2 = 3 4 · 1 2 · 1 2 + 3 4 · 1 2 · 1 2

  12. Inferring the Causal Direction ◮ We observe X = x , then we observe Y = y . ◮ What is the probability of H = h ? ◮ Calculate posterior probability: P ( y | h , x ) P ( x | h ) P ( h ) P ( h | x , y ) = P ( y | h , x ) P ( x | h ) P ( h ) + P ( x |¬ h , y ) P ( y |¬ h ) P ( ¬ h ) 4 · 1 3 2 · 1 = 1 2 = 2 = P ( h )! 3 4 · 1 2 · 1 2 + 3 4 · 1 2 · 1 2

  13. Inferring the Causal Direction ◮ We observe X = x , then we observe Y = y . ◮ What is the probability of H = h ? ◮ Calculate posterior probability: P ( y | h , x ) P ( x | h ) P ( h ) P ( h | x , y ) = P ( y | h , x ) P ( x | h ) P ( h ) + P ( x |¬ h , y ) P ( y |¬ h ) P ( ¬ h ) 3 4 · 1 2 · 1 = 1 2 = 2 = P ( h )! 3 4 · 1 2 · 1 2 + 3 4 · 1 2 · 1 2 ◮ We haven’t learned anything!

  14. Inferring the Causal Direction ◮ We observe X = x , then we observe Y = y . ◮ What is the probability of H = h ? ◮ Calculate posterior probability: P ( y | h , x ) P ( x | h ) P ( h ) P ( h | x , y ) = P ( y | h , x ) P ( x | h ) P ( h ) + P ( x |¬ h , y ) P ( y |¬ h ) P ( ¬ h ) 3 4 · 1 2 · 1 = 1 2 = 2 = P ( h )! 4 · 1 3 2 · 1 2 + 3 4 · 1 2 · 1 2 ◮ We haven’t learned anything! ◮ To extract new causal information, we have to supply old causal information: ◮ “no causes in, no causes out” ◮ “to learn what happens if you kick the system, you have to kick the system”

  15. Interventions in a Probability Tree Set X = x : H 1 1 2 2 � �� X Y � �� � �� 1 1 1 1 2 2 2 2 Y Y X X � �� � �� � �� � �� 3 1 1 3 3 1 1 3 4 4 4 4 4 4 4 4 3 1 1 3 3 1 1 3 P ( X , Y | H ) : 8 8 8 8 8 8 8 8

  16. Interventions in a Probability Tree Set X = x : H 1 1 2 2 � �� X Y � �� � �� 1 1 1 0 2 2 Y Y X X � �� � �� � �� � �� 3 1 1 3 1 0 1 0 4 4 4 4 3 1 1 1 P ( X , Y | H ) : 0 0 0 0 4 4 2 2 ◮ Replace all mechanisms resolving X with the delta “ X = x ”.

  17. Inferring the Causal Direction—2nd Attempt ◮ We set X = x , then we observe Y = y . ◮ What is the probability of H = h ?

  18. Inferring the Causal Direction—2nd Attempt ◮ We set X = x , then we observe Y = y . ◮ What is the probability of H = h ? ◮ Calculate posterior probability: P ( y | h , ˆ x ) P (ˆ x | h ) P ( h ) P ( h | ˆ x , y ) = P ( y | h , ˆ x ) P (ˆ x | h ) P ( h ) + P (ˆ x |¬ h , y ) P ( y |¬ h ) P ( ¬ h ) 3 4 · 1 · 1 2 = 3 4 · 1 · 1 2 + 1 · 1 2 · 1 2

  19. Inferring the Causal Direction—2nd Attempt ◮ We set X = x , then we observe Y = y . ◮ What is the probability of H = h ? ◮ Calculate posterior probability: P ( y | h , ˆ x ) P (ˆ x | h ) P ( h ) P ( h | ˆ x , y ) = P ( y | h , ˆ x ) P (ˆ x | h ) P ( h ) + P (ˆ x |¬ h , y ) P ( y |¬ h ) P ( ¬ h ) 3 4 · 1 · 1 = 3 2 = 5 � = P ( h ) . 3 4 · 1 · 1 2 + 1 · 1 2 · 1 2 ◮ We have have acquired evidence for “ X → Y ”!

  20. Conclusions ◮ Causal induction can be done using purely Bayesian techniques plus a description allowing multiple causal explanations of an experiment. ◮ Probability trees provide a clean & simple way to encode causal probabilistic information. ◮ The purpose of an intervention is to introduce statistical asymmetries. ◮ The causal information that we can acquire is limited by the interventions we can apply to the system. ◮ In this approach, the causal dependencies are not “in the data”, but they rather arise from the data and the hypotheses that the reasoner “imprints” on them.

Download Presentation
Download Policy: The content available on the website is offered to you 'AS IS' for your personal information and use only. It cannot be commercialized, licensed, or distributed on other websites without prior consent from the author. To download a presentation, simply click this link. If you encounter any difficulties during the download process, it's possible that the publisher has removed the file from their server.

Recommend


More recommend