Causal Inference and Graphical Models - II
Jin Tian Iowa State University
– p.1
Causal Inference and Graphical Models - II Jin Tian Iowa State - - PowerPoint PPT Presentation
Causal Inference and Graphical Models - II Jin Tian Iowa State University p.1 Outline Computing the effects of manipulations Inferring constraints implied by DAGs with hidden variables on nonexperimental data on experimental data
Jin Tian Iowa State University
– p.1
– p.2
Cancer Smoking Tar in lungs Z Y U X
– p.3
P(U) P(X|U) P(Y|Z,U) Tar in lungs Cancer Smoking P(Z|X) Z Y X U Tar in do(X=False) P(U) lungs Cancer Smoking P(Z|X) P(Y|Z,U) Z Y U X
– p.4
t(s) = P(s|do(t)) = P(s|set(t)) = P(s|ˆ
– p.5
i
t(v) = ∏ {i|Vi∈T}
– p.6
X Y U
– p.7
Unidentifiable
X Y U
u
u
x (y) =∑ u
x (y) =∑ u
x (y) = PM2 x (y)
– p.8
X Z Y U
– p.9
X Z Y U
z
x′
– p.9
– p.10
X Z Y U
z
z
Rule 2
z
Rule 2
z
Rule 3
z ∑ x′
z
x′
– p.11
X Z Y U X Y U
– p.12
X Z Y U X Z Y U X Y U X Y U
– p.12
2 1
U U Z1 Z Y X 2
– p.13
u
{i|Vi∈V}
{i|Ui∈U}
For any set S ⊆ V, define
u
{i|Vi∈S}
{i|Ui∈U}
– p.14
u
{i|Vi∈V}
{i|Ui∈U}
For any set S ⊆ V, define
u
{i|Vi∈S}
{i|Ui∈U}
partitioned into c-components S1,...,Sk. Then
i
i
– p.14
2 1
U U Z1 Z Y X 2
Two c-components:
u1,u2
– p.15
2 1
U U Z1 Z Y X 2
Two c-components:
u1,u2
u1
u2
– p.15
{j|Vj∈Si}
– p.16
– p.17
2 1
U U Z1 Z Y X 2
– p.18
2 1
U U Z1 Z Y X 2
– p.18
2 1
U U Z1 Z Y X 2
– p.18
i
i
– p.19
– p.20
t(s) = ∑ (v\t)\s
t(v\t)
(v\t)\s
d\s∏ i
t(s) is identifiable iff each Q[Di] is identifiable.
– p.21
t(s) to be unidentifiable or
t(s) in terms of P(v)
– p.22
– p.23
– p.24
– p.24
b
b
b
– p.25
A B C D U A B C D U
Independence statements: A is independent of C given B.
– p.26
A B C D U
Consider
u
b
Therefore ∑b P(d|a,b,c)P(b|a) is independent of a.
– p.27
– p.28
u
{i|Vi∈S}
{i|Ui∈U}
– p.29
– p.30
U V V V V
1
U2
2 1 4 3
– p.31
Pearl’s instrumental inequality, for discrete variables
x ∑ y
z
E.g., binary variables
– p.32
– p.33
A causal BN not only imposes constraints on the nonexperimental distribution but also on the experimental distributions A causal BN can be tested and falsified by using two types
nonexperimental data are passively observed, experimental data are produced by manipulating (randomly) some variables and observing the states of
The ability to use a mixture of nonexperimental and experimental data will greatly increase our power of causal reasoning and learning.
– p.34
i
– p.35
– p.36
S2⊆Ti\S1
– p.37
– p.38
– p.39
Equality constraints: Pz(xy) = P(xy|z), Pxz(y) = Px(y) Inequality: Pz(xy) ≤ Pxz(y) We have
z
y
z
– p.40
W1 X U1 Y Z U2 W2 U3
yz
w1 P(z|w1xw2y)P(y|w1xw2)P(x|w1) ≤ 1.
– p.41
t(v)|T ⊆ V,t ∈ Dm(T)}
– p.42
t(t) = 1.
– p.43
– p.44
S2⊆V\S1
– p.45
– p.46
– p.47
– p.48
– p.49
– p.50
– p.51
– p.52
– p.53
{u | Y(u)=y}
{u | Yx(u)=y}
{u|Yx(u)=y & X(u)=x′}
{u | Yx(u)=y & Yx′(u)=y′}
– p.54
– p.55
where U1 and U2 are two independent binary variables with
2, leading to the same distribution
Model 1: P(Yx=0 = 0|X = 1,Y = 1) = 0 Model 2: P(Yx=0 = 0|X = 1,Y = 1) = 1
– p.56
– p.57
Let X and Y be two binary variables Probability of necessity (PN)
x′|x,y)
PN stands for the probability that event y would not have
x′, given that x and y
did in fact occur. Applications in epidemiology, legal reasoning, and AI: a certain case of disease is attributable to a particular exposure, “the probability that disease would not have
exposure did in fact occur.”
– p.58
– p.59
– p.60
– p.61
Parameters: p110 = P(yx,yx′,x′),... Probabilistic constraints:
1
i=0 1
j=0 1
k=0
Nonexperimental constraints:
– p.62
– p.63
P(x,y)
P(y′
x′)−P(x′,y′)
P(x,y)
P(x′,y′)
P(yx)−P(x,y) P(x′,y′)
– p.65
– p.66
– p.67