Bias/Variance Analysis for Network Data Jennifer Neville and David - - PowerPoint PPT Presentation

bias variance analysis for network data
SMART_READER_LITE
LIVE PREVIEW

Bias/Variance Analysis for Network Data Jennifer Neville and David - - PowerPoint PPT Presentation

Bias/Variance Analysis for Network Data Jennifer Neville and David Jensen Knowledge Discovery Laboratory Knowledge Discovery Laboratory University of Massachusetts Amherst University of Massachusetts Amherst Collective inference + + + + +


slide-1
SLIDE 1

Jennifer Neville and David Jensen

Knowledge Discovery Laboratory Knowledge Discovery Laboratory University of Massachusetts Amherst University of Massachusetts Amherst

Bias/Variance Analysis for Network Data

slide-2
SLIDE 2

2/13

+ + + + – + + – – + – – – – – – – – – – – – – – – + + + – – – + + + – + + – – + + + – + + – –

Collective inference

− Apply models to collectively infer class labels throughout network − Exploit autocorrelation to improve model performance − Collective SRL models

− Probabilistic relational models (e.g., RBNs, RDNs, RMNs) − Probabilistic logic models (e.g., BLPs, MLNs) − Adhoc collective models (e.g., pRNs, LBC)

+ + +

+

– + +

– + – – – – – – – – – – – – –

– + + + –

– + + + – +

+

– –

+

+ + – + + – – + + +

+

– + +

– + – – – – – – – – – – – – –

– + + + –

– + + + – +

+

– –

+

+ + – + + – –

slide-3
SLIDE 3

3/13

Comparing collective models

Relational dependency networks Latent group models

slide-4
SLIDE 4

3/13

Comparing collective models

Relational dependency networks Latent group models

Why do RDNs perform poorly when few instances are labeled in test set?

slide-5
SLIDE 5

4/13

Understanding RDN performance − Hypothesis

− High autocorrelation → features selection chooses class label rather than observed attributes − Few labeled test set instances → identifiability problem − Gibbs sampling → increased variance

− How to evaluate hypothesis?

− Variance is due to collective inference procedure − Need an analysis framework that can differentiate model errors due to learning and inference

slide-6
SLIDE 6

5/13

Bias/variance analysis − Conventional bias/variance analysis

− Decomposes errors due to learning alone − Assumes no variation due to inference

− Relational bias/variance analysis

− Collective inference introduces new source of error − SRL models exhibit different types of errors − Network characteristics affect performance

slide-7
SLIDE 7

6/13

Conventional bias/variance framework

Training Set Samples

M1 M2 M3

Models Test Set Model predictions

slide-8
SLIDE 8

6/13

Conventional bias/variance framework

Training Set Samples

M1 M2 M3

Models Test Set Model predictions

bias variance

Y* Y _

−Expected Expected error error per per instance instance −Decompose Decompose into into model model bias/variance bias/variance

slide-9
SLIDE 9

7/13

Bias/variance framework for relational data

Training Set Samples

M1 M2 M3

Models Fully labeled Test Set

+ + + + – + + – – + – – – – – – – – – – – – – – – + + + – – – – – – – – – – – – – + + – – + + + – – + + + – + + – – – – – – – – – – – –

Model predictions

slide-10
SLIDE 10

7/13

Bias/variance framework for relational data

Training Set Samples

M1 M2 M3

Models Fully labeled Test Set

+ + + + – + + – – + – – – – – – – – – – – – – – – + + + – – – – – – – – – – – – – + + – – + + + – – + + + – + + – – – – – – – – – – – –

Model predictions Y*

−Measure Measure learning bias bias and and variance variance with with full full labeling labeling

learning bias

YL _

learning variance

slide-11
SLIDE 11

8/13

– – – – – – – – – – – – – – – – – – – –

Bias/variance framework for relational data

Training Set Samples

M1 M2 M3

Models Test Set Inference Runs

+ + + + – + + – – + – – – – – – – – – – – – – – – + + + – – – – – – – – – – – – – + + – – + + + – – + + + – + + – –

Model predictions

– – – – – – – – – – – – – – – – – – – –

slide-12
SLIDE 12

8/13

– – – – – – – – – – – – – – – – – – – –

Bias/variance framework for relational data

Training Set Samples

M1 M2 M3

Models Test Set Inference Runs

+ + + + – + + – – + – – – – – – – – – – – – – – – + + + – – – – – – – – – – – – – + + – – + + + – – + + + – + + – –

Model predictions

– – – – – – – – – – – – – – – – – – – –

Y* Y _

−Measure Measure total bias bias and and variance variance

−Expectation over training Expectation over training and test sets test sets

total bias total variance

slide-13
SLIDE 13

8/13

– – – – – – – – – – – – – – – – – – – –

Bias/variance framework for relational data

Training Set Samples

M1 M2 M3

Models Test Set Inference Runs

+ + + + – + + – – + – – – – – – – – – – – – – – – + + + – – – – – – – – – – – – – + + – – + + + – – + + + – + + – –

Model predictions

– – – – – – – – – – – – – – – – – – – –

Y* Y _

−Measure Measure total bias bias and and variance variance

−Expectation over training Expectation over training and test sets test sets

total bias total variance

Y* Y _

−Measure Measure learning bias bias and and variance variance with with full full labeling labeling −Measure Measure total bias bias and and variance variance

−Expectation over training Expectation over training and test sets test sets

−Difference: Difference: inference bias bias and and variance variance

YL _

inference bias learning bias

slide-14
SLIDE 14

9/13

Synthetic data experiments − Vary group size, linkage, autocorrelation − Compare LGMs, RDNs, RMNs − Preliminary findings

− LGMs: high learning bias when algorithm cannot identify underlying group structure − RDNs: high inference variance when little information seeding inference process − RMNs: high inference bias when network is densely connected or tightly clustered

slide-15
SLIDE 15

10/13

Feature selection increases RDN inference variance

slide-16
SLIDE 16

10/13

Feature selection increases RDN inference variance

Inference Variance

slide-17
SLIDE 17

11/13

Modified inference decreases variance

slide-18
SLIDE 18

12/13

Improved performance on real data

slide-19
SLIDE 19

13/13

Conclusions − Framework can be used to explain mechanisms behind SRL model performance

− Improves understanding of model behavior − Suggests algorithmic modifications to increase performance

− Future work

− Extend framework (e.g., loss functions, joint estimation) − Investigate interaction effects between learning and inference errors − Real data experiments to evaluate design choices

slide-20
SLIDE 20

14/13

Further information:

jneville@cs.umass.edu kdl.cs.umass.edu