Quantum Causal Structures
Christian Majenz
University of Copenhagen Joint work with Rafael Chaves and David Gross, University of Freiburg (arXiv:1407.3800)
04.12.2014
Quantum Causal Structures Christian Majenz University of Copenhagen - - PowerPoint PPT Presentation
Quantum Causal Structures Christian Majenz University of Copenhagen Joint work with Rafael Chaves and David Gross, University of Freiburg (arXiv:1407.3800) 04.12.2014 Motivation Question Question Can we rule out certain causal relationships
Christian Majenz
University of Copenhagen Joint work with Rafael Chaves and David Gross, University of Freiburg (arXiv:1407.3800)
04.12.2014
Can we rule out certain causal relationships by just looking at statistical data?
Can we rule out certain causal relationships by just looking at statistical data? Example:
Smoking Lung Cancer Gene
unobserved
Can we rule out certain causal relationships by just looking at statistical data? Example:
Smoking Lung Cancer Gene
unobserved
◮ Rigorous mathematical way to distinguish different causal
relations
◮ Rigorous mathematical way to distinguish different causal
relations
◮ Introduced by Pearl in the 1980s
◮ Rigorous mathematical way to distinguish different causal
relations
◮ Introduced by Pearl in the 1980s ◮ Necessary conditions for probability distributions to come from
more restrictive causal model
◮ Rigorous mathematical way to distinguish different causal
relations
◮ Introduced by Pearl in the 1980s ◮ Necessary conditions for probability distributions to come from
more restrictive causal model
◮ Tools: Bayesian networks, entropies, convexity
◮ Rigorous mathematical way to distinguish different causal
relations
◮ Introduced by Pearl in the 1980s ◮ Necessary conditions for probability distributions to come from
more restrictive causal model
◮ Tools: Bayesian networks, entropies, convexity
[?] What if the underlying processes are quantum?
◮ Rigorous mathematical way to distinguish different causal
relations
◮ Introduced by Pearl in the 1980s ◮ Necessary conditions for probability distributions to come from
more restrictive causal model
◮ Tools: Bayesian networks, entropies, convexity
[?] What if the underlying processes are quantum?
◮ Bell nonlocality etc. special cases of this
Motivation Classical Bayesian Networks Entropic Description Quantum Quantum Causal Structure Entropic Description Application Information Causality Computational Techniques
◮ Any directed acyclic graph (DAG) specifies a causal structure:
1 2 3 4 5 6
◮ Any directed acyclic graph (DAG) specifies a causal structure:
1 2 3 4 5 6
◮ Directed graph: G = (V , E) with E ⊂ V × V
◮ Any directed acyclic graph (DAG) specifies a causal structure:
1 2 3 4 5 6
◮ Directed graph: G = (V , E) with E ⊂ V × V ◮ Acyclic: no cycle (causality)
◮ Any directed acyclic graph (DAG) specifies a causal structure:
1 2 3 4 5 6
◮ Directed graph: G = (V , E) with E ⊂ V × V ◮ Acyclic: no cycle (causality) ◮ parents of v: pa(v) = {w ∈ V |(w, v) ∈ E}
◮ Any directed acyclic graph (DAG) specifies a causal structure:
1 2 3 4 5 6
◮ Directed graph: G = (V , E) with E ⊂ V × V ◮ Acyclic: no cycle (causality) ◮ parents of v: pa(v) = {w ∈ V |(w, v) ∈ E} ◮ children, ancestors, descendants, non-descendants etc.
Definition (Bayesian network)
Let G = (V , E) be a DAG, X = (Xv)v∈V a collection of random variables indexed by V with probability distribution p on X V . Then p is called Markov w.r.t. G if p(x1, ..., xn) =
p
assuming WLOG V = {1, ..., n}. In that case G together with X is called a Bayesian network.
◮ equivalent characterization: v⊥nd(v)|pa(v)
◮ equivalent characterization: v⊥nd(v)|pa(v) ◮ example:
X → Y → Z = ⇒ Z⊥X|Y ⇔ p(x, y, z)p(y) = p(x, y)p(y, z)
◮ equivalent characterization: v⊥nd(v)|pa(v) ◮ example:
X → Y → Z = ⇒ Z⊥X|Y ⇔ p(x, y, z)p(y) = p(x, y)p(y, z)
◮ formalism requires access to joint probability distribution
◮ equivalent characterization: v⊥nd(v)|pa(v) ◮ example:
X → Y → Z = ⇒ Z⊥X|Y ⇔ p(x, y, z)p(y) = p(x, y)p(y, z)
◮ formalism requires access to joint probability distribution
= ⇒ obstacle for quantum generalization
◮ equivalent characterization: v⊥nd(v)|pa(v) ◮ example:
X → Y → Z = ⇒ Z⊥X|Y ⇔ p(x, y, z)p(y) = p(x, y)p(y, z)
◮ formalism requires access to joint probability distribution
= ⇒ obstacle for quantum generalization
◮ hard algebraic equations
X random variable (RV) on a finite alphabet X probability distribution p
X random variable (RV) on a finite alphabet X probability distribution p Shannon entropy: H(X) = −
p(x) log p(x)
X random variable (RV) on a finite alphabet X probability distribution p Shannon entropy: H(X) = −
p(x) log p(x) ρ ∈ S(H) mixed state on a finite-dimensional Hilbert space H = Cd
X random variable (RV) on a finite alphabet X probability distribution p Shannon entropy: H(X) = −
p(x) log p(x) ρ ∈ S(H) mixed state on a finite-dimensional Hilbert space H = Cd Von Neumann entropy S(ρ) = −trρ log ρ
Classical:
◮ Collection of RV’s X = (X1, ..., Xn)
Classical:
◮ Collection of RV’s X = (X1, ..., Xn)
XI, I ⊂ {1, ..., n}
Classical:
◮ Collection of RV’s X = (X1, ..., Xn)
XI, I ⊂ {1, ..., n}
◮ Shannon entropy vector: h(X) = (H(XI))I⊂{1,...,n} ∈ R2n
Classical:
◮ Collection of RV’s X = (X1, ..., Xn)
XI, I ⊂ {1, ..., n}
◮ Shannon entropy vector: h(X) = (H(XI))I⊂{1,...,n} ∈ R2n
Quantum:
◮ Multipartite state ρ ∈ S(H) with H = H1 ⊗ ... ⊗ Hn
Classical:
◮ Collection of RV’s X = (X1, ..., Xn)
XI, I ⊂ {1, ..., n}
◮ Shannon entropy vector: h(X) = (H(XI))I⊂{1,...,n} ∈ R2n
Quantum:
◮ Multipartite state ρ ∈ S(H) with H = H1 ⊗ ... ⊗ Hn
ρI = trIc ρ
Classical:
◮ Collection of RV’s X = (X1, ..., Xn)
XI, I ⊂ {1, ..., n}
◮ Shannon entropy vector: h(X) = (H(XI))I⊂{1,...,n} ∈ R2n
Quantum:
◮ Multipartite state ρ ∈ S(H) with H = H1 ⊗ ... ⊗ Hn
ρI = trIc ρ
◮ Von Neumann entropy vector: s(ρ) = (S(ρI))I⊂{1,...,n} ∈ R2n
What is a convex cone?
What is a convex cone?
What is a convex cone?
Sets of all entropy vectors:
Sets of all entropy vectors: Classical Σn = {h(X)|X n-party random variable}
Sets of all entropy vectors: Classical Σn = {h(X)|X n-party random variable} Quantum Γn = {s(ρ)|ρ n-party quantum state}
Sets of all entropy vectors: Classical Σn = {h(X)|X n-party random variable} Quantum Γn = {s(ρ)|ρ n-party quantum state} Σn as well as Γn are convex cones (Zhang, Yeung, 1997 and Pippenger, 2003)
Describe convex cone via linear inequalities
Describe convex cone via linear inequalities
Describe convex cone via linear inequalities Inequalities for Σn: information inequalities
Describe convex cone via linear inequalities Inequalities for Σn: information inequalities Example: monotonicity H(X1X2) − H(X2) ≥ 0
Describe convex cone via linear inequalities Inequalities for Σn: information inequalities Example: monotonicity H(X1X2) − H(X2) ≥ 0 Inequalities for Γn: quantum information inequalities
Describe convex cone via linear inequalities Inequalities for Σn: information inequalities Example: monotonicity H(X1X2) − H(X2) ≥ 0 Inequalities for Γn: quantum information inequalities Example: weak monotonicity S(ρ12) + S(ρ13) − S(ρ2) − S(ρ3) ≥ 0
Describe convex cone via linear inequalities Inequalities for Σn: information inequalities Example: monotonicity H(X1X2) − H(X2) ≥ 0 Inequalities for Γn: quantum information inequalities Example: weak monotonicity S(ρ12) + S(ρ13) − S(ρ2) − S(ρ3) ≥ 0 known (quantum) information inequalities provide outer approximation of entropy cones
◮ Consider entropy cone of RVs from Bayesian network
◮ Consider entropy cone of RVs from Bayesian network ◮ Conditonal independence: Z⊥X|Y ⇔ I(X : Z|Y ) = 0
◮ Consider entropy cone of RVs from Bayesian network ◮ Conditonal independence: Z⊥X|Y ⇔ I(X : Z|Y ) = 0 ◮ Conditional Mutual information
I(X : Z|Y ) = H(XY ) + H(YZ) − H(XZY ) − H(Y )
◮ Consider entropy cone of RVs from Bayesian network ◮ Conditonal independence: Z⊥X|Y ⇔ I(X : Z|Y ) = 0 ◮ Conditional Mutual information
I(X : Z|Y ) = H(XY ) + H(YZ) − H(XZY ) − H(Y )
◮ Linear equation
◮ Consider entropy cone of RVs from Bayesian network ◮ Conditonal independence: Z⊥X|Y ⇔ I(X : Z|Y ) = 0 ◮ Conditional Mutual information
I(X : Z|Y ) = H(XY ) + H(YZ) − H(XZY ) − H(Y )
◮ Linear equation ◮ intersection with entropy cone is convex cone
◮ Only some RVs are observed
◮ Only some RVs are observed ◮ Example:
◮ Only some RVs are observed ◮ Example:
→ Marginal scenario: {A, B, C}
◮ Only some RVs are observed ◮ Example:
→ Marginal scenario: {A, B, C}
◮ Plan: Remove unobserved variables from inequality
description of entropy cone
Definition (Marginal Scenario)
Let n ∈ N. A subset M ⊂ 2{1,...,n} such that for I ∈ M and J ⊂ I also J ∈ M is called marginal scenario.
Definition (Marginal Scenario)
Let n ∈ N. A subset M ⊂ 2{1,...,n} such that for I ∈ M and J ⊂ I also J ∈ M is called marginal scenario. Projection of entropy cone to a marginal scenario = ⇒ observable entropy cone.
◮ DAG G = (V , E)
◮ DAG G = (V , E) ◮ Sink nodes s ∈ V : Hilbert space Hs
◮ DAG G = (V , E) ◮ Sink nodes s ∈ V : Hilbert space Hs ◮ Nodes v ∈ V with children: Hilbert space
Hv =
(v,w)∈E
Hv,w
◮ DAG G = (V , E) ◮ Sink nodes s ∈ V : Hilbert space Hs ◮ Nodes v ∈ V with children: Hilbert space
Hv =
(v,w)∈E
Hv,w
◮
H =
Hv
◮ DAG G = (V , E) ◮ Sink nodes s ∈ V : Hilbert space Hs ◮ Nodes v ∈ V with children: Hilbert space
Hv =
(v,w)∈E
Hv,w
◮
H =
Hv
◮ Parent Hilbert space
Hpa(v) =
(w,v)∈E
Hw,v
◮ Initial product state ρ0 = q ρq on Hilbert spaces of source
nodes q
◮ Initial product state ρ0 = q ρq on Hilbert spaces of source
nodes q
◮ CPTP maps Φv for each non-source node:
Φv : L
◮ Initial product state ρ0 = q ρq on Hilbert spaces of source
nodes q
◮ CPTP maps Φv for each non-source node:
Φv : L
◮
ρv = Φvρpa(v)
◮ Initial product state ρ0 = q ρq on Hilbert spaces of source
nodes q
◮ CPTP maps Φv for each non-source node:
Φv : L
◮
ρv = Φvρpa(v) [!] no global state
◮ Initial product state ρ0 = q ρq on Hilbert spaces of source
nodes q
◮ CPTP maps Φv for each non-source node:
Φv : L
◮
ρv = Φvρpa(v) [!] no global state
◮ want classical nodes: pick the right Φv
2 3 1
unobserved
◮ H = H1,2 ⊗ H1,3 ⊗ H2 ⊗ H3
2 3 1
unobserved
◮ H = H1,2 ⊗ H1,3 ⊗ H2 ⊗ H3 ◮ States on the coexisting subsets of systems:
ρ0 = ρ(1,2),(1,3) ρ(1,3),2 = (Φ2 ⊗ 1) ρ(1,2),(1,3) ρ(1,2),3 = (1 ⊗ Φ3) ρ(1,2),(1,3) ρ2,3 = (Φ2 ⊗ Φ3) ρ(1,2),(1,3)
◮ Look at entropy cone of states constructed like this
◮ Look at entropy cone of states constructed like this ◮ H is a tensor product of n = |E| + |Vs| hilbert spaces, Vs: Set
◮ Look at entropy cone of states constructed like this ◮ H is a tensor product of n = |E| + |Vs| hilbert spaces, Vs: Set
◮ Formally: Entropy vector v ∈ R2n
◮ Look at entropy cone of states constructed like this ◮ H is a tensor product of n = |E| + |Vs| hilbert spaces, Vs: Set
◮ Formally: Entropy vector v ∈ R2n ◮ vI = S(ρI) if the state ρI exists
◮ Look at entropy cone of states constructed like this ◮ H is a tensor product of n = |E| + |Vs| hilbert spaces, Vs: Set
◮ Formally: Entropy vector v ∈ R2n ◮ vI = S(ρI) if the state ρI exists ◮ vI arbitrary if ρI doesn’t exist
◮ Look at entropy cone of states constructed like this ◮ H is a tensor product of n = |E| + |Vs| hilbert spaces, Vs: Set
◮ Formally: Entropy vector v ∈ R2n ◮ vI = S(ρI) if the state ρI exists ◮ vI arbitrary if ρI doesn’t exist ◮ for each J ⊂ E ∪ Vs of coexisting systems: quantum entropy
cone
◮ Look at entropy cone of states constructed like this ◮ H is a tensor product of n = |E| + |Vs| hilbert spaces, Vs: Set
◮ Formally: Entropy vector v ∈ R2n ◮ vI = S(ρI) if the state ρI exists ◮ vI arbitrary if ρI doesn’t exist ◮ for each J ⊂ E ∪ Vs of coexisting systems: quantum entropy
cone
◮ extra monotonicities for classical systems
◮ replace conditional independence relations?
◮ replace conditional independence relations?
→ Data processing inequality: For a CPTP map Φ : HA → HB I(A : C|D) ≥ I(B : C|D) for any other systems C and D.
◮ replace conditional independence relations?
→ Data processing inequality: For a CPTP map Φ : HA → HB I(A : C|D) ≥ I(B : C|D) for any other systems C and D.
◮ relates entropies of noncoexisting systems
◮ replace conditional independence relations?
→ Data processing inequality: For a CPTP map Φ : HA → HB I(A : C|D) ≥ I(B : C|D) for any other systems C and D.
◮ relates entropies of noncoexisting systems ◮ replaces conditional independence
◮ Again, only some systems are observed
◮ Again, only some systems are observed ◮ Example:
◮ Again, only some systems are observed ◮ Example:
◮ Again, only some systems are observed ◮ Example:
◮ Again, only some systems are observed ◮ Example: ◮ most interesting marginal scenario: {A, B, C} & assume these
nodes classical
The IC principle is defined by a game:
The IC principle is defined by a game:
The IC principle is defined by a game:
The IC principle is defined by a game:
ΡAB
The IC principle is defined by a game:
X1,X2
ΡAB
The IC principle is defined by a game:
M
X1,X2
ΡAB
The IC principle is defined by a game:
M
X1,X2
i
ΡAB
The IC principle is defined by a game:
M
X1,X2
i
ΡAB Yi
The IC principle is defined by a game:
M
X1,X2
i
ΡAB Yi
◮ Corresponding DAG: X1 S X2 Y M AB
◮ Corresponding DAG: X1 S X2 Y M AB ◮ Nodes Xi, Y , M, S are classical, only the AB node is quantum
◮ Corresponding DAG: X1 S X2 Y M AB ◮ Nodes Xi, Y , M, S are classical, only the AB node is quantum ◮ Counterfactual reformulation: Yi = (Y |S = i)
◮ Marginal scenario {X1, Y1}, {X2, Y2}, {M} yields original
information causality inequality I(X1 : Y1) + I(X2 : Y2) ≤ H(M)
◮ Marginal scenario {X1, Y1}, {X2, Y2}, {M} yields original
information causality inequality I(X1 : Y1) + I(X2 : Y2) ≤ H(M) → More general marginal scenario {X1, X2, Y1, M}, {X1, X2, Y2, M}:
◮ Marginal scenario {X1, Y1}, {X2, Y2}, {M} yields original
information causality inequality I(X1 : Y1) + I(X2 : Y2) ≤ H(M) → More general marginal scenario {X1, X2, Y1, M}, {X1, X2, Y2, M}:
X1 S X2 Y M AB
◮ Marginal scenario {X1, Y1}, {X2, Y2}, {M} yields original
information causality inequality I(X1 : Y1) + I(X2 : Y2) ≤ H(M) → More general marginal scenario {X1, X2, Y1, M}, {X1, X2, Y2, M}: Strengthened information causality inequality I(X1 : Y1, M)+I(X2 : Y2, M)+I(X1 : X2|M) ≤ H(M)+I(X1 : X2)
1
◮ Classical: e.g.
I(A : B) + I(A : C) ≤ H(A)
1
◮ Classical: e.g.
I(A : B) + I(A : C) ≤ H(A)
◮ New framework =
⇒ also holds for quantum
1
◮ Classical: e.g.
I(A : B) + I(A : C) ≤ H(A)
◮ New framework =
⇒ also holds for quantum
◮ independently proven by Henson et al., 2014 for non-signalling
resources1
1Henson, Joe, Raymond Lal, and Matthew F. Pusey. ”Theory-independent
limits on correlations from generalised Bayesian networks.” arXiv preprint arXiv:1405.2572 (2014).
◮ Classical: e.g.
I(A : B) + I(A : C) ≤ H(A)
◮ New framework =
⇒ also holds for quantum
◮ independently proven by Henson et al., 2014 for non-signalling
resources1
◮ using their techniques =
⇒ causes connecting at most m of n nodes yield entropic inequality
1Henson, Joe, Raymond Lal, and Matthew F. Pusey. ”Theory-independent
limits on correlations from generalised Bayesian networks.” arXiv preprint arXiv:1405.2572 (2014).
◮ Fancy name for removing variables from inequalities
◮ Fancy name for removing variables from inequalities ◮ Example: x ≤ 2y, z ≤ x + y, y ≤ x, remove y
◮ Fancy name for removing variables from inequalities ◮ Example: x ≤ 2y, z ≤ x + y, y ≤ x, remove y
= ⇒
1 2x ≤ x, z − x ≤ x
◮ Fancy name for removing variables from inequalities ◮ Example: x ≤ 2y, z ≤ x + y, y ≤ x, remove y
= ⇒
1 2x ≤ x, z − x ≤ x ◮ Problem: double exponential in number of variables...
◮ Fancy name for removing variables from inequalities ◮ Example: x ≤ 2y, z ≤ x + y, y ≤ x, remove y
= ⇒
1 2x ≤ x, z − x ≤ x ◮ Problem: double exponential in number of variables...
... i.e. triple exponential in number of nodes
◮ Fancy name for removing variables from inequalities ◮ Example: x ≤ 2y, z ≤ x + y, y ≤ x, remove y
= ⇒
1 2x ≤ x, z − x ≤ x ◮ Problem: double exponential in number of variables...
... i.e. triple exponential in number of nodes
◮ works for small instances (triangle, IC)
◮ Easier: check candidate inequality
◮ Easier: check candidate inequality ◮ formulation as LP:
minimize
αIvI subject to SSA, WM, DPs
◮ Easier: check candidate inequality ◮ formulation as LP:
minimize
αIvI subject to SSA, WM, DPs
◮ minimum is 0 if the inequality holds
◮ Easier: check candidate inequality ◮ formulation as LP:
minimize
αIvI subject to SSA, WM, DPs
◮ minimum is 0 if the inequality holds ◮ −∞ else
◮ Easier: check candidate inequality ◮ formulation as LP:
minimize
αIvI subject to SSA, WM, DPs
◮ minimum is 0 if the inequality holds ◮ −∞ else
[!] still exponential in # of nodes
◮ Defined quantum causal structures
◮ Defined quantum causal structures ◮ Algorithm for characterization
◮ Defined quantum causal structures ◮ Algorithm for characterization ◮ Applications: strengthening of information causality, quantum
networks
Thank you for your attention!