Dynamic Bayesian Networks Cora Perez-Ariza 1 , Ann Nicholson 2 , - - PowerPoint PPT Presentation

dynamic bayesian networks
SMART_READER_LITE
LIVE PREVIEW

Dynamic Bayesian Networks Cora Perez-Ariza 1 , Ann Nicholson 2 , - - PowerPoint PPT Presentation

Causal Discovery of Dynamic Bayesian Networks Cora Perez-Ariza 1 , Ann Nicholson 2 , Kevin Korb 2 , Steven Mascaro 2 and Chao Heng Hu 2 1 Dept. of Computer Science and Artificial Intelligence University of Granada 2 Clayton School of IT Monash


slide-1
SLIDE 1

Causal Discovery of Dynamic Bayesian Networks

Cora Perez-Ariza1, Ann Nicholson2, Kevin Korb2, Steven Mascaro2 and Chao Heng Hu2

1

  • Dept. of Computer Science and Artificial Intelligence

University of Granada

2

Clayton School of IT Monash University

slide-2
SLIDE 2

Learning Static BNs

Constraint-based learning

 Performs independence tests  e.g. PC algorithm (Spirtes et al., 1993)  Tests all pairs for direct dependencies

slide-3
SLIDE 3

Learning Static BNs

Constraint-based learning

 Performs independence tests  e.g. PC algorithm (Spirtes et al., 1993)  Tests all pairs for direct dependencies Finding a graph pattern Finding head to head arcs and orient the rest

slide-4
SLIDE 4

Learning Static BNs

Metric-based learning

 (Stochastic) search and score  Learning programs/packages: e.g. CaMML (Causal discovery via

MML), BNT (Bayes Net Toolbox).

slide-5
SLIDE 5

Learning Static BNs

Metric-based learning

 (Stochastic) search and score  Learning programs/packages: e.g. CaMML (Causal discovery via

MML), BNT (Bayes Net Toolbox).

Score and rank M → M' : Add/remove/reverse arcs add remove reverse

slide-6
SLIDE 6

t2

Dynamic Bayesian Networks

Extension of BN with arcs from t → t + 1 DBNs that we consider:

  • 1. Same structure for each slice (i.e. stationary)
  • 2. Arcs cannot span more than one time step

t0 t1

slide-7
SLIDE 7

Learning Dynamic Bayesian Networks

Why not use existing static learners?

Need to guarantee slice t nodes come before slice t+1 nodes

Often want slices to be the same (i.e. stationary)

Make the search more efficient  Produce better models

slide-8
SLIDE 8

Learning DBNs – Previous Approaches

Friedman et al. (1998)

 Uses BIC/BDe scoring

 Hill-climbing  Learn the prior/initial network and the transition network

Prior network Transition network

+

The corresponding DBN

slide-9
SLIDE 9

Learning DBNs – Previous Approaches

Bayes Net Toolbox (BNT)

 Written by Kevin Murphy (2001)  Supports DBN learning and inference

BNT algorithm

 Uses BIC/ML scoring  Guarantees that  Only learns arcs between slices (temporal arcs)

slide-10
SLIDE 10

Two New Approaches to Learning DBNs

  • 1. Enforce stationary DBN structure with structural

priors

  • 2. Enhance existing search and score procedure to

take DBN structure into account

Both take advantage of our BN learner software CaMML

slide-11
SLIDE 11

CaMML

Bayesian network learner created at Monash

Uses MML for score and MCMC for search

Can specify flexible priors:

  • A -> B: Direct causal connection
  • A – B: Direct relation
  • A => B: Ancestral relation
  • A ~ B: Correlation
  • Tiers
  • Existing BN structure
slide-12
SLIDE 12

CaMML

Bayesian network learner created at Monash

Uses MML for score and MCMC for search

Can specify flexible priors:

  • A -> B: Direct causal connection
  • A – B: Direct relation
  • A => B: Ancestral relation
  • A ~ B: Correlation
  • Tiers
  • Existing BN structure
slide-13
SLIDE 13

CaMML Tier Priors Learning

slide-14
SLIDE 14

CaMML Tier Priors Learning

t2 t0 t1

Motivation, SES, Education ≺ Motivation1, SES1, Education1

slide-15
SLIDE 15

CaMML 2-Step Learning

Learn transitional arcs t0 Data BN for t0 Learn from Data Learned DBN Copy Network t0 t1

slide-16
SLIDE 16

Experiments

Test models

We compare CaMML against two other learning programs:

  • PC algorithm (in GeNIe)
  • BNT
slide-17
SLIDE 17

Milk Infection DBN

Mutual information Use mutual information to score strength of arcs

t0 t1

slide-18
SLIDE 18

BAT DBN

Too many variables to show!

t0 t1

slide-19
SLIDE 19

Plain static BN learning (without using priors)

CaMML vs GeNIe (PC algorithm)

Experiment #1

Learning with tier priors

CaMML vs GeNIe and BNT

Experiment #3 Experiment #2

Learning using tier priors vs 2-step algorithm

slide-20
SLIDE 20

Experiment Procedure

Known Models Tier or BN Priors Learned Models Generate Data Learn DBNs Test with ED/CKL

slide-21
SLIDE 21

Evaluation

Edit distance

Count 1 if an arc is missing/added/reversed in the learned model Our modification for DBNs: EDDBN = Ws.Ns + Wt.Nt

Causal Kullback-Leibler divergence

Computes the distance of probability distribution between model P and model Q

slide-22
SLIDE 22

Milk and cancer model (CaMML vs GeNIe, using tiers)

Results

slide-23
SLIDE 23

Milk and cancer model (CaMML vs GeNIe, using tiers)

Results

slide-24
SLIDE 24

Results

Transitional arc errors for the BAT network

Datasize CaMML w/ Tiers GeNIe (PC) BNT 500 6.8 (0.98) 13 7.5 (0.50) 5000 5.4 (1.62) 10 6.2 (0.74) 50000 1.0 (0.0) 10 3.7 (0.33)

slide-25
SLIDE 25

5 10 15 20 25 30 35 total errors (tiers) total errors (Alg)

Results

BAT model (CaMML, tiers vs 2-step learning)

500 5000 50000 Data size Errors

slide-26
SLIDE 26

5 10 15 20 25 static errors (tiers) static errors (Alg)

BAT model (CaMML, tiers vs 2-step learning)

Errors

Results

500 5000 50000 Data size

slide-27
SLIDE 27

2 4 6 8 10 12 14 temporal errors (tiers) temporal errors (Alg)

BAT model (CaMML, tiers vs 2-step learning)

500 5000 50000 Data size Errors

Results

slide-28
SLIDE 28

Summary

GeNIe(PC) tends to over-fit (i.e. more arcs added) with large data size in Experiment #1. Using tiers, CaMML produces fewer errors than BNT and GeNIe(PC). CaMML can recover more weak arcs, and usually learns all the strong arcs. The 2-step learning algorithm produces comparable results, better at learning static arcs.

slide-29
SLIDE 29

CaMML 2-Step Learning Issues

slide-30
SLIDE 30

CaMML 2-Step Learning Issues

slide-31
SLIDE 31

Current and Future Work

 Modify CaMML’s search and score:

  • Alter score to avoid double counting static

arcs

  • Alter search to avoid invalid DBN structures
  • Ultimately: Reduce the search space so that

we can find good models more quickly