Phylogenetic Support 1. Introduction 2. Evolutionary Model Testing - - PowerPoint PPT Presentation

phylogenetic support
SMART_READER_LITE
LIVE PREVIEW

Phylogenetic Support 1. Introduction 2. Evolutionary Model Testing - - PowerPoint PPT Presentation

Finlay Maguire Statistical Testing of Trees March 27, 2018 FCS, Dalhousie Phylogenetic Support 1. Introduction 2. Evolutionary Model Testing 3. Branch Support Testing 4. Comparing Trees 5. Conclusion 1 Table of contents Introduction 2


slide-1
SLIDE 1

Phylogenetic Support

Statistical Testing of Trees

Finlay Maguire March 27, 2018

FCS, Dalhousie

slide-2
SLIDE 2

Table of contents

  • 1. Introduction
  • 2. Evolutionary Model Testing
  • 3. Branch Support Testing
  • 4. Comparing Trees
  • 5. Conclusion

1

slide-3
SLIDE 3

Introduction

slide-4
SLIDE 4

Phylogenies are hypotheses

Cid

Tetrahymena thermophila [XP_001012854.1] Tetrahymena thermophila [XP_001012858.1] Paramecium bursaria SW1 [comp3906_seq0_m.68533] Paramecium bursaria SW1 [comp3906_seq0_m.68531] Paramecium bursaria Yad1g [TR17851_c0_g1_i8_m.235761] Paramecium bursaria Yad1g [TR432_c1_g1_i2_m.4057]

80.7%/0.93

Paramecium biaurelia [PBIGNP33303] Paramecium tetaurelia Cid3 [GSPATP00025353001] Paramecium sexaurelia [PSEXPNG26288]

89.8%/0.94

Paramecium multimicronucleatum [PMMNP07604]

99.8%/1.00

Paramecium caudatum [PCAUDP10462]

91%/0.93

Paramecium tetaurelia Cid1 (Marker, 2014) [PTETP9100013001] Paramecium biaurelia [PBIGNP26212] Paramecium primaurelia [PPRIMP23072]

5%/0.51

Paramecium sexaurelia [PSEXPNG26738]

42%/0.71

Paramecium multimicronucleatum [PMMNP02964]

98.9%/0.99

Paramecium caudatum [PCAUDP15935]

55.4%/0.63 99.7%/1.00 59.5%/0.67 100%/1.00 97.9%/1.00

Paramecium caudatum [PSEXPNG26858] Paramecium multimicronucleatum [PMMNP03007] Paramecium sexaurelia [PSEXPNG26858] Paramecium primaurelia [PPRIMP27560] Paramecium biaurelia [PBIGNP11073] Paramecium tetaurelia Cid2 (Marker, 2014) [PTETP13400003001]

84.1%/0.91 83%/0.88 95.3%/0.96 83.9%/0.88 59.1%/0.54 99.7%/1.00 86.7%/0.69 100%/1.00

0.2 Cid2 Cid1 Cid3 Cid1-3 Ancestor?

2

slide-5
SLIDE 5

Hypothesis testing

  • Does another model of sequence evolution fit the data better?
  • How well supported are individual branches in a tree?
  • Does another tree explain the data better?

3

slide-6
SLIDE 6

Sources of Error

  • Bad data
  • Sampling error
  • Misleading evolutionary events
  • Misspecified models
  • Inappropriate inference

4

slide-7
SLIDE 7

Sources of Error

  • Bad data
  • Sampling error
  • Misleading evolutionary events
  • Misspecified models
  • Inappropriate inference

4

slide-8
SLIDE 8

Sources of Error

  • Bad data
  • Sampling error
  • Misleading evolutionary events
  • Misspecified models
  • Inappropriate inference

4

slide-9
SLIDE 9

Sources of Error

  • Bad data
  • Sampling error
  • Misleading evolutionary events
  • Misspecified models
  • Inappropriate inference

4

slide-10
SLIDE 10

Sources of Error

  • Bad data
  • Sampling error
  • Misleading evolutionary events
  • Misspecified models
  • Inappropriate inference

4

slide-11
SLIDE 11

Saturation

[Leonard, 2010]

5

slide-12
SLIDE 12

Misleading Signal: Recombination

6

slide-13
SLIDE 13

Misleading Signal: Hidden Paralogy/Incomplete Sampling

[Leonard, 2010]

7

slide-14
SLIDE 14

Misleading Signal: Horizontal Gene Transfer

[Leonard, 2010]

8

slide-15
SLIDE 15

Misleading Signal: Horizontal Gene Transfer

[Richards et al., 2009]

9

slide-16
SLIDE 16

Tree not always correct paradigm

Ask for a tree get a tree.

10

slide-17
SLIDE 17

Tree not always correct paradigm

Ask for a tree get a tree.

Reanalysis of [Marwick, 2012] from http://phylonetworks.blogspot.ca/2013/02/

11

slide-18
SLIDE 18

Evolutionary Model Testing

slide-19
SLIDE 19

Sequence Evolution Models

http: //carrot.mcb.uconn.edu/~olgazh/bioinf2010/class24.html

12

slide-20
SLIDE 20

Sequence Evolution Models

[Nickle et al., 2007]

13

slide-21
SLIDE 21

What happens if the wrong model is specified?

  • Increased Inaccuracy (wrong tree more often)
  • Inconsistency (adding more data converges to wrong tree)
  • Wrong branch lengths (important for certain analyses)
  • Wrong tree support values

14

slide-22
SLIDE 22

What happens if the wrong model is specified?

  • Increased Inaccuracy (wrong tree more often)
  • Inconsistency (adding more data converges to wrong tree)
  • Wrong branch lengths (important for certain analyses)
  • Wrong tree support values

14

slide-23
SLIDE 23

What happens if the wrong model is specified?

  • Increased Inaccuracy (wrong tree more often)
  • Inconsistency (adding more data converges to wrong tree)
  • Wrong branch lengths (important for certain analyses)
  • Wrong tree support values

14

slide-24
SLIDE 24

What happens if the wrong model is specified?

  • Increased Inaccuracy (wrong tree more often)
  • Inconsistency (adding more data converges to wrong tree)
  • Wrong branch lengths (important for certain analyses)
  • Wrong tree support values

14

slide-25
SLIDE 25

How do we select a model?

  • L(τ, θ) = P(X|τ, θ)
  • With ML inference we are finding the maximum-likelihood

estimate of and

  • i.e.

argmax L

  • Therefore, to compare two models we can use a likelihood ratio

test (LRT )

  • 2 ln L1

ln L0

  • Limitations: nested models (i.e. hLRT), order matters, no

regularisation

15

slide-26
SLIDE 26

How do we select a model?

  • L(τ, θ) = P(X|τ, θ)
  • With ML inference we are finding the maximum-likelihood

estimate of τ and θ

  • i.e.

argmax L

  • Therefore, to compare two models we can use a likelihood ratio

test (LRT )

  • 2 ln L1

ln L0

  • Limitations: nested models (i.e. hLRT), order matters, no

regularisation

15

slide-27
SLIDE 27

How do we select a model?

  • L(τ, θ) = P(X|τ, θ)
  • With ML inference we are finding the maximum-likelihood

estimate of τ and θ

  • i.e. ˆ

τ, ˆ θ = argmaxτ,θ L(τ, θ)

  • Therefore, to compare two models we can use a likelihood ratio

test (LRT )

  • 2 ln L1

ln L0

  • Limitations: nested models (i.e. hLRT), order matters, no

regularisation

15

slide-28
SLIDE 28

How do we select a model?

  • L(τ, θ) = P(X|τ, θ)
  • With ML inference we are finding the maximum-likelihood

estimate of τ and θ

  • i.e. ˆ

τ, ˆ θ = argmaxτ,θ L(τ, θ)

  • Therefore, to compare two models we can use a likelihood ratio

test (LRT δ)

  • 2 ln L1

ln L0

  • Limitations: nested models (i.e. hLRT), order matters, no

regularisation

15

slide-29
SLIDE 29

How do we select a model?

  • L(τ, θ) = P(X|τ, θ)
  • With ML inference we are finding the maximum-likelihood

estimate of τ and θ

  • i.e. ˆ

τ, ˆ θ = argmaxτ,θ L(τ, θ)

  • Therefore, to compare two models we can use a likelihood ratio

test (LRT δ)

  • δ = 2(ln(L1) − ln(L0))
  • Limitations: nested models (i.e. hLRT), order matters, no

regularisation

15

slide-30
SLIDE 30

How do we select a model?

  • L(τ, θ) = P(X|τ, θ)
  • With ML inference we are finding the maximum-likelihood

estimate of τ and θ

  • i.e. ˆ

τ, ˆ θ = argmaxτ,θ L(τ, θ)

  • Therefore, to compare two models we can use a likelihood ratio

test (LRT δ)

  • δ = 2(ln(L1) − ln(L0))
  • Limitations: nested models (i.e. hLRT), order matters, no

regularisation

15

slide-31
SLIDE 31

Information Criterion

  • Akaike Information Criterion (AIC), penalising number of

parameters:

  • AIC

2ln L 2K

  • However, this penalises all high K models even if sample size is

large too.

  • Corrected Akaike Information Criterion (AICc)
  • AICc

AIC

2K K 1 n K 1

  • Alternatively, there is the Bayesian Information Criterion (BIC):
  • BIC

2ln L Kln n

  • Decision Theory (DT) risk minimisation approach.

16

slide-32
SLIDE 32

Information Criterion

  • Akaike Information Criterion (AIC), penalising number of

parameters:

  • AIC = −2ln(L) + 2K
  • However, this penalises all high K models even if sample size is

large too.

  • Corrected Akaike Information Criterion (AICc)
  • AICc

AIC

2K K 1 n K 1

  • Alternatively, there is the Bayesian Information Criterion (BIC):
  • BIC

2ln L Kln n

  • Decision Theory (DT) risk minimisation approach.

16

slide-33
SLIDE 33

Information Criterion

  • Akaike Information Criterion (AIC), penalising number of

parameters:

  • AIC = −2ln(L) + 2K
  • However, this penalises all high K models even if sample size is

large too.

  • Corrected Akaike Information Criterion (AICc)
  • AICc

AIC

2K K 1 n K 1

  • Alternatively, there is the Bayesian Information Criterion (BIC):
  • BIC

2ln L Kln n

  • Decision Theory (DT) risk minimisation approach.

16

slide-34
SLIDE 34

Information Criterion

  • Akaike Information Criterion (AIC), penalising number of

parameters:

  • AIC = −2ln(L) + 2K
  • However, this penalises all high K models even if sample size is

large too.

  • Corrected Akaike Information Criterion (AICc)
  • AICc

AIC

2K K 1 n K 1

  • Alternatively, there is the Bayesian Information Criterion (BIC):
  • BIC

2ln L Kln n

  • Decision Theory (DT) risk minimisation approach.

16

slide-35
SLIDE 35

Information Criterion

  • Akaike Information Criterion (AIC), penalising number of

parameters:

  • AIC = −2ln(L) + 2K
  • However, this penalises all high K models even if sample size is

large too.

  • Corrected Akaike Information Criterion (AICc)
  • AICc = AIC + 2K(K+1)

n−K−1

  • Alternatively, there is the Bayesian Information Criterion (BIC):
  • BIC

2ln L Kln n

  • Decision Theory (DT) risk minimisation approach.

16

slide-36
SLIDE 36

Information Criterion

  • Akaike Information Criterion (AIC), penalising number of

parameters:

  • AIC = −2ln(L) + 2K
  • However, this penalises all high K models even if sample size is

large too.

  • Corrected Akaike Information Criterion (AICc)
  • AICc = AIC + 2K(K+1)

n−K−1

  • Alternatively, there is the Bayesian Information Criterion (BIC):
  • BIC

2ln L Kln n

  • Decision Theory (DT) risk minimisation approach.

16

slide-37
SLIDE 37

Information Criterion

  • Akaike Information Criterion (AIC), penalising number of

parameters:

  • AIC = −2ln(L) + 2K
  • However, this penalises all high K models even if sample size is

large too.

  • Corrected Akaike Information Criterion (AICc)
  • AICc = AIC + 2K(K+1)

n−K−1

  • Alternatively, there is the Bayesian Information Criterion (BIC):
  • BIC = −2ln(L) + Kln(n)
  • Decision Theory (DT) risk minimisation approach.

16

slide-38
SLIDE 38

Information Criterion

  • Akaike Information Criterion (AIC), penalising number of

parameters:

  • AIC = −2ln(L) + 2K
  • However, this penalises all high K models even if sample size is

large too.

  • Corrected Akaike Information Criterion (AICc)
  • AICc = AIC + 2K(K+1)

n−K−1

  • Alternatively, there is the Bayesian Information Criterion (BIC):
  • BIC = −2ln(L) + Kln(n)
  • Decision Theory (DT) risk minimisation approach.

16

slide-39
SLIDE 39

Limitations

  • What if everything fits poorly?
  • Information criterion test relative goodness of fit instead of

absolute

  • Parametric Bootstrapping/Posterior Predictive Simulation
  • If the model is reasonable then data simulated under should

resemble the empirical data

17

slide-40
SLIDE 40

Limitations

  • What if everything fits poorly?
  • Information criterion test relative goodness of fit instead of

absolute

  • Parametric Bootstrapping/Posterior Predictive Simulation
  • If the model is reasonable then data simulated under should

resemble the empirical data

17

slide-41
SLIDE 41

Limitations

  • What if everything fits poorly?
  • Information criterion test relative goodness of fit instead of

absolute

  • Parametric Bootstrapping/Posterior Predictive Simulation
  • If the model is reasonable then data simulated under should

resemble the empirical data

17

slide-42
SLIDE 42

Limitations

  • What if everything fits poorly?
  • Information criterion test relative goodness of fit instead of

absolute

  • Parametric Bootstrapping/Posterior Predictive Simulation
  • If the model is reasonable then data simulated under should

resemble the empirical data

17

slide-43
SLIDE 43

Branch Support Testing

slide-44
SLIDE 44

Bootstrapping in General

The bootstrap

(unknown) true value of (unknown) true distribution empirical distribution of sample estimate of Distribution of estimates

  • f parameters

Bootstrap replicates

Slide from Joe Felsenstein

18

slide-45
SLIDE 45

Bootstrapping Phylogenies

The bootstrap for phylogenies

Original Data sites Bootstrap sample #1 Bootstrap sample #2

sample same number

  • f sites, with replacement

sample same number

  • f sites, with replacement

(and so on)

T

^

T(1) T(2)

Slide from Joe Felsenstein

19

slide-46
SLIDE 46

Bootstrapping Phylogenies

20

slide-47
SLIDE 47

Bootstrapping Phylogenies

The majority-rule consensus tree

C A

Trees: How many times each partition of species is found: AE | BCDF 4 ACE | BDF 3 ACEF | BD 1 AC | BDEF 1 AEF | BCD 1 ADEF | BC 2 ABCE | DF 3

B D F E C A B D F E C A B D F E C A B D F E C A B D F E A C

B D F E

0.6 0.6 0.8

Slide from Joe Felsenstein

21

slide-48
SLIDE 48

Combining the results

22

slide-49
SLIDE 49

What is the bootstrap doing?

  • Randomly reweighing the sites in an alignments
  • Probability of a site being excluded 1

1 nn

  • Asymptotically approximately 0 36
  • Goal to simulate an infinite population (number of alignment

columns)

23

slide-50
SLIDE 50

What is the bootstrap doing?

  • Randomly reweighing the sites in an alignments
  • Probability of a site being excluded 1 − 1

nn

  • Asymptotically approximately 0 36
  • Goal to simulate an infinite population (number of alignment

columns)

23

slide-51
SLIDE 51

What is the bootstrap doing?

  • Randomly reweighing the sites in an alignments
  • Probability of a site being excluded 1 − 1

nn

  • Asymptotically approximately 0.36
  • Goal to simulate an infinite population (number of alignment

columns)

23

slide-52
SLIDE 52

What is the bootstrap doing?

  • Randomly reweighing the sites in an alignments
  • Probability of a site being excluded 1 − 1

nn

  • Asymptotically approximately 0.36
  • Goal to simulate an infinite population (number of alignment

columns)

23

slide-53
SLIDE 53

Limitations

  • Typically underestimates the true probabilities
  • i.e biased but conservative
  • Computationally demanding
  • Assumes independence of sites
  • Relies on good input data
  • Only answers to what extent does input data support a given

part of the tree

24

slide-54
SLIDE 54

Limitations

  • Typically underestimates the true probabilities
  • i.e biased but conservative
  • Computationally demanding
  • Assumes independence of sites
  • Relies on good input data
  • Only answers to what extent does input data support a given

part of the tree

24

slide-55
SLIDE 55

Limitations

  • Typically underestimates the true probabilities
  • i.e biased but conservative
  • Computationally demanding
  • Assumes independence of sites
  • Relies on good input data
  • Only answers to what extent does input data support a given

part of the tree

24

slide-56
SLIDE 56

Limitations

  • Typically underestimates the true probabilities
  • i.e biased but conservative
  • Computationally demanding
  • Assumes independence of sites
  • Relies on good input data
  • Only answers to what extent does input data support a given

part of the tree

24

slide-57
SLIDE 57

Limitations

  • Typically underestimates the true probabilities
  • i.e biased but conservative
  • Computationally demanding
  • Assumes independence of sites
  • Relies on good input data
  • Only answers to what extent does input data support a given

part of the tree

24

slide-58
SLIDE 58

Limitations

  • Typically underestimates the true probabilities
  • i.e biased but conservative
  • Computationally demanding
  • Assumes independence of sites
  • Relies on good input data
  • Only answers to what extent does input data support a given

part of the tree

24

slide-59
SLIDE 59

Parametric Bootstraps

  • Simulate data sets of this size assuming the estimate of the tree

is the truth

  • Key for many more sophisticated tests.
  • Can be used to generate p-values, but non-trivial

25

slide-60
SLIDE 60

Alternative Approaches

  • Resampling estimated log-likelihoods (RELL)
  • Instead of re-doing the full ML inference just re-sample the site

ln L values and sum

  • Rapid Bootstraps (RBS)
  • Ultrafast Bootstraps (UFBoot)

26

slide-61
SLIDE 61

Alternative Approaches

  • Resampling estimated log-likelihoods (RELL)
  • Instead of re-doing the full ML inference just re-sample the site

ln(L) values and sum

  • Rapid Bootstraps (RBS)
  • Ultrafast Bootstraps (UFBoot)

26

slide-62
SLIDE 62

Alternative Approaches

  • Resampling estimated log-likelihoods (RELL)
  • Instead of re-doing the full ML inference just re-sample the site

ln(L) values and sum

  • Rapid Bootstraps (RBS)
  • Ultrafast Bootstraps (UFBoot)

26

slide-63
SLIDE 63

Alternative Approaches

  • Resampling estimated log-likelihoods (RELL)
  • Instead of re-doing the full ML inference just re-sample the site

ln(L) values and sum

  • Rapid Bootstraps (RBS)
  • Ultrafast Bootstraps (UFBoot)

26

slide-64
SLIDE 64

Likelihood Tests

  • Comparing the 3 nearest NNIs

to a given branch:

  • Parametric aLRT:

2 of

for branch vs. closest NNIs

  • Non-parametric SH-aLRT

based on RELL

  • aBayes:
  • P Tc

X

P X Tc P Tc

2 i

0P X Ti P Ti with

flat prior P T0 P T1 P T2

27

slide-65
SLIDE 65

Likelihood Tests

  • Comparing the 3 nearest NNIs

to a given branch:

  • Parametric aLRT: χ2 of δ for

branch vs. closest NNIs

  • Non-parametric SH-aLRT

based on RELL

  • aBayes:
  • P Tc

X

P X Tc P Tc

2 i

0P X Ti P Ti with

flat prior P T0 P T1 P T2

27

slide-66
SLIDE 66

Likelihood Tests

  • Comparing the 3 nearest NNIs

to a given branch:

  • Parametric aLRT: χ2 of δ for

branch vs. closest NNIs

  • Non-parametric SH-aLRT

based on RELL

  • aBayes:
  • P Tc

X

P X Tc P Tc

2 i

0P X Ti P Ti with

flat prior P T0 P T1 P T2

27

slide-67
SLIDE 67

Likelihood Tests

  • Comparing the 3 nearest NNIs

to a given branch:

  • Parametric aLRT: χ2 of δ for

branch vs. closest NNIs

  • Non-parametric SH-aLRT

based on RELL

  • aBayes:
  • P Tc

X

P X Tc P Tc

2 i

0P X Ti P Ti with

flat prior P T0 P T1 P T2

27

slide-68
SLIDE 68

Likelihood Tests

  • Comparing the 3 nearest NNIs

to a given branch:

  • Parametric aLRT: χ2 of δ for

branch vs. closest NNIs

  • Non-parametric SH-aLRT

based on RELL

  • aBayes:
  • P(Tc | X) =

P(X|Tc)P(Tc) ∑2

i =0P(X||Ti)P(Ti) with

flat prior P(T0) = P(T1) = P(T2)

27

slide-69
SLIDE 69

Comparing Trees

slide-70
SLIDE 70

How to compare competing hypotheses?

https://github.com/mtholder/TreeTopoTestingTalks

28

slide-71
SLIDE 71

How to compare competing hypotheses?

https://github.com/mtholder/TreeTopoTestingTalks

29

slide-72
SLIDE 72

Simplistic Comparison

30

slide-73
SLIDE 73

Qualitative Comparison

  • 4 sites favour the red tree, 2 favour the blue
  • n

k pk 1

p n

k

  • 4 out of 6 p

0 6875

  • 40 out of 60 p

0 0124

  • 400 out of 600 p

2 3 10

16 31

slide-74
SLIDE 74

Qualitative Comparison

  • 4 sites favour the red tree, 2 favour the blue
  • (n

k

) pk(1 − p)n−k

  • 4 out of 6 p

0 6875

  • 40 out of 60 p

0 0124

  • 400 out of 600 p

2 3 10

16 31

slide-75
SLIDE 75

Qualitative Comparison

  • 4 sites favour the red tree, 2 favour the blue
  • (n

k

) pk(1 − p)n−k

  • 4 out of 6 p = 0.6875
  • 40 out of 60 p

0 0124

  • 400 out of 600 p

2 3 10

16 31

slide-76
SLIDE 76

Qualitative Comparison

  • 4 sites favour the red tree, 2 favour the blue
  • (n

k

) pk(1 − p)n−k

  • 4 out of 6 p = 0.6875
  • 40 out of 60 p = 0.0124
  • 400 out of 600 p

2 3 10

16 31

slide-77
SLIDE 77

Qualitative Comparison

  • 4 sites favour the red tree, 2 favour the blue
  • (n

k

) pk(1 − p)n−k

  • 4 out of 6 p = 0.6875
  • 40 out of 60 p = 0.0124
  • 400 out of 600 p = 2.3 ∗ 10−16

31

slide-78
SLIDE 78

Quantiative Comparison

  • µ = (−5.2 + 3.1 + 0.9 + 6.6 + 0.3 − 0.2)/6 = 0.916
  • 2

15 22

  • t

2

N 0 148

  • therefore: p

0 888 under 5d f

32

slide-79
SLIDE 79

Quantiative Comparison

  • µ = (−5.2 + 3.1 + 0.9 + 6.6 + 0.3 − 0.2)/6 = 0.916
  • σ2 = 15.22
  • t

2

N 0 148

  • therefore: p

0 888 under 5d f

32

slide-80
SLIDE 80

Quantiative Comparison

  • µ = (−5.2 + 3.1 + 0.9 + 6.6 + 0.3 − 0.2)/6 = 0.916
  • σ2 = 15.22
  • t = µ

σ2 ∗

√ N = 0.148

  • therefore: p

0 888 under 5d f

32

slide-81
SLIDE 81

Quantiative Comparison

  • µ = (−5.2 + 3.1 + 0.9 + 6.6 + 0.3 − 0.2)/6 = 0.916
  • σ2 = 15.22
  • t = µ

σ2 ∗

√ N = 0.148

  • therefore: p = 0.888 under 5d.f.

32

slide-82
SLIDE 82

More robust approaches

  • Null: if no sampling error (infinite data) T1 and T2 would explain

the data equally well.

  • T1 T2

X 2 ln L T1 X ln L T2 X

  • Expectation under null

T1 T2 X

  • Why can’t we just use

2 to get a critical value for ?

  • Tree space is difficult.

33

slide-83
SLIDE 83

More robust approaches

  • Null: if no sampling error (infinite data) T1 and T2 would explain

the data equally well.

  • δ(T1, T2 | X) = 2 [ln L(T1 | X) − ln L(T2 | X)]
  • Expectation under null

T1 T2 X

  • Why can’t we just use

2 to get a critical value for ?

  • Tree space is difficult.

33

slide-84
SLIDE 84

More robust approaches

  • Null: if no sampling error (infinite data) T1 and T2 would explain

the data equally well.

  • δ(T1, T2 | X) = 2 [ln L(T1 | X) − ln L(T2 | X)]
  • Expectation under null E [δ(T1, T2 | X)] = 0
  • Why can’t we just use

2 to get a critical value for ?

  • Tree space is difficult.

33

slide-85
SLIDE 85

More robust approaches

  • Null: if no sampling error (infinite data) T1 and T2 would explain

the data equally well.

  • δ(T1, T2 | X) = 2 [ln L(T1 | X) − ln L(T2 | X)]
  • Expectation under null E [δ(T1, T2 | X)] = 0
  • Why can’t we just use χ2 to get a critical value for δ?
  • Tree space is difficult.

33

slide-86
SLIDE 86

More robust approaches

  • Null: if no sampling error (infinite data) T1 and T2 would explain

the data equally well.

  • δ(T1, T2 | X) = 2 [ln L(T1 | X) − ln L(T2 | X)]
  • Expectation under null E [δ(T1, T2 | X)] = 0
  • Why can’t we just use χ2 to get a critical value for δ?
  • Tree space is difficult.

33

slide-87
SLIDE 87

Estimating variance of the null

  • Many avenues:
  • Non-parametric bootstrapping
  • Parametric bootstrapping
  • Related approaches.

34

slide-88
SLIDE 88

Kishino-Hasegawa Test

  • First, winning sites test
  • H0

ln L T1 X ln L T2 X T1 T2

  • Ha

T1 T2

  • Non-parametric Bootstrap to estimate Null variance
  • Test

T1 T2 two-tail t-test

  • Due to centring assumption can’t be used for optimal tree i.e.

selection bias

  • Can’t handle multiple comparisons.

35

slide-89
SLIDE 89

Kishino-Hasegawa Test

  • First, winning sites test
  • H0 : [ln L(T1 | X) − ln L(T2 | X)] = E [δ(T1, T2)] = 0
  • Ha

T1 T2

  • Non-parametric Bootstrap to estimate Null variance
  • Test

T1 T2 two-tail t-test

  • Due to centring assumption can’t be used for optimal tree i.e.

selection bias

  • Can’t handle multiple comparisons.

35

slide-90
SLIDE 90

Kishino-Hasegawa Test

  • First, winning sites test
  • H0 : [ln L(T1 | X) − ln L(T2 | X)] = E [δ(T1, T2)] = 0
  • Ha : E [δ(T1, T2)] ̸= 0
  • Non-parametric Bootstrap to estimate Null variance
  • Test

T1 T2 two-tail t-test

  • Due to centring assumption can’t be used for optimal tree i.e.

selection bias

  • Can’t handle multiple comparisons.

35

slide-91
SLIDE 91

Kishino-Hasegawa Test

  • First, winning sites test
  • H0 : [ln L(T1 | X) − ln L(T2 | X)] = E [δ(T1, T2)] = 0
  • Ha : E [δ(T1, T2)] ̸= 0
  • Non-parametric Bootstrap to estimate Null variance
  • Test

T1 T2 two-tail t-test

  • Due to centring assumption can’t be used for optimal tree i.e.

selection bias

  • Can’t handle multiple comparisons.

35

slide-92
SLIDE 92

Kishino-Hasegawa Test

  • First, winning sites test
  • H0 : [ln L(T1 | X) − ln L(T2 | X)] = E [δ(T1, T2)] = 0
  • Ha : E [δ(T1, T2)] ̸= 0
  • Non-parametric Bootstrap to estimate Null variance
  • Test E [δ(T1, T2)] two-tail t-test
  • Due to centring assumption can’t be used for optimal tree i.e.

selection bias

  • Can’t handle multiple comparisons.

35

slide-93
SLIDE 93

Kishino-Hasegawa Test

  • First, winning sites test
  • H0 : [ln L(T1 | X) − ln L(T2 | X)] = E [δ(T1, T2)] = 0
  • Ha : E [δ(T1, T2)] ̸= 0
  • Non-parametric Bootstrap to estimate Null variance
  • Test E [δ(T1, T2)] two-tail t-test
  • Due to centring assumption can’t be used for optimal tree i.e.

selection bias

  • Can’t handle multiple comparisons.

35

slide-94
SLIDE 94

Kishino-Hasegawa Test

  • First, winning sites test
  • H0 : [ln L(T1 | X) − ln L(T2 | X)] = E [δ(T1, T2)] = 0
  • Ha : E [δ(T1, T2)] ̸= 0
  • Non-parametric Bootstrap to estimate Null variance
  • Test E [δ(T1, T2)] two-tail t-test
  • Due to centring assumption can’t be used for optimal tree i.e.

selection bias

  • Can’t handle multiple comparisons.

35

slide-95
SLIDE 95

Alternative tests

  • Shimodaira-Hasegawa Test
  • Compares candidate tree sets
  • H0

all topologies equally good

  • Very conservative when the number of candidate trees is large
  • Can be corrected with weighted SH-test overcomes.
  • Approximately Unbiased Test
  • Achieves weighted by varying bootstrap size for each tree.
  • Better for larger comparisons, can have issues with P-space

curvature.

  • Swofford–Olsen–Waddell–Hillis same idea but uses parametric

bootstraps instead.

  • Sensitive to model misspecification.

36

slide-96
SLIDE 96

Alternative tests

  • Shimodaira-Hasegawa Test
  • Compares candidate tree sets
  • H0

all topologies equally good

  • Very conservative when the number of candidate trees is large
  • Can be corrected with weighted SH-test overcomes.
  • Approximately Unbiased Test
  • Achieves weighted by varying bootstrap size for each tree.
  • Better for larger comparisons, can have issues with P-space

curvature.

  • Swofford–Olsen–Waddell–Hillis same idea but uses parametric

bootstraps instead.

  • Sensitive to model misspecification.

36

slide-97
SLIDE 97

Alternative tests

  • Shimodaira-Hasegawa Test
  • Compares candidate tree sets
  • H0 = all topologies equally good
  • Very conservative when the number of candidate trees is large
  • Can be corrected with weighted SH-test overcomes.
  • Approximately Unbiased Test
  • Achieves weighted by varying bootstrap size for each tree.
  • Better for larger comparisons, can have issues with P-space

curvature.

  • Swofford–Olsen–Waddell–Hillis same idea but uses parametric

bootstraps instead.

  • Sensitive to model misspecification.

36

slide-98
SLIDE 98

Alternative tests

  • Shimodaira-Hasegawa Test
  • Compares candidate tree sets
  • H0 = all topologies equally good
  • Very conservative when the number of candidate trees is large
  • Can be corrected with weighted SH-test overcomes.
  • Approximately Unbiased Test
  • Achieves weighted by varying bootstrap size for each tree.
  • Better for larger comparisons, can have issues with P-space

curvature.

  • Swofford–Olsen–Waddell–Hillis same idea but uses parametric

bootstraps instead.

  • Sensitive to model misspecification.

36

slide-99
SLIDE 99

Alternative tests

  • Shimodaira-Hasegawa Test
  • Compares candidate tree sets
  • H0 = all topologies equally good
  • Very conservative when the number of candidate trees is large
  • Can be corrected with weighted SH-test overcomes.
  • Approximately Unbiased Test
  • Achieves weighted by varying bootstrap size for each tree.
  • Better for larger comparisons, can have issues with P-space

curvature.

  • Swofford–Olsen–Waddell–Hillis same idea but uses parametric

bootstraps instead.

  • Sensitive to model misspecification.

36

slide-100
SLIDE 100

Alternative tests

  • Shimodaira-Hasegawa Test
  • Compares candidate tree sets
  • H0 = all topologies equally good
  • Very conservative when the number of candidate trees is large
  • Can be corrected with weighted SH-test overcomes.
  • Approximately Unbiased Test
  • Achieves weighted by varying bootstrap size for each tree.
  • Better for larger comparisons, can have issues with P-space

curvature.

  • Swofford–Olsen–Waddell–Hillis same idea but uses parametric

bootstraps instead.

  • Sensitive to model misspecification.

36

slide-101
SLIDE 101

Alternative tests

  • Shimodaira-Hasegawa Test
  • Compares candidate tree sets
  • H0 = all topologies equally good
  • Very conservative when the number of candidate trees is large
  • Can be corrected with weighted SH-test overcomes.
  • Approximately Unbiased Test
  • Achieves weighted by varying bootstrap size for each tree.
  • Better for larger comparisons, can have issues with P-space

curvature.

  • Swofford–Olsen–Waddell–Hillis same idea but uses parametric

bootstraps instead.

  • Sensitive to model misspecification.

36

slide-102
SLIDE 102

Alternative tests

  • Shimodaira-Hasegawa Test
  • Compares candidate tree sets
  • H0 = all topologies equally good
  • Very conservative when the number of candidate trees is large
  • Can be corrected with weighted SH-test overcomes.
  • Approximately Unbiased Test
  • Achieves weighted by varying bootstrap size for each tree.
  • Better for larger comparisons, can have issues with P-space

curvature.

  • Swofford–Olsen–Waddell–Hillis same idea but uses parametric

bootstraps instead.

  • Sensitive to model misspecification.

36

slide-103
SLIDE 103

Alternative tests

  • Shimodaira-Hasegawa Test
  • Compares candidate tree sets
  • H0 = all topologies equally good
  • Very conservative when the number of candidate trees is large
  • Can be corrected with weighted SH-test overcomes.
  • Approximately Unbiased Test
  • Achieves weighted by varying bootstrap size for each tree.
  • Better for larger comparisons, can have issues with P-space

curvature.

  • Swofford–Olsen–Waddell–Hillis same idea but uses parametric

bootstraps instead.

  • Sensitive to model misspecification.

36

slide-104
SLIDE 104

Alternative tests

  • Shimodaira-Hasegawa Test
  • Compares candidate tree sets
  • H0 = all topologies equally good
  • Very conservative when the number of candidate trees is large
  • Can be corrected with weighted SH-test overcomes.
  • Approximately Unbiased Test
  • Achieves weighted by varying bootstrap size for each tree.
  • Better for larger comparisons, can have issues with P-space

curvature.

  • Swofford–Olsen–Waddell–Hillis same idea but uses parametric

bootstraps instead.

  • Sensitive to model misspecification.

36

slide-105
SLIDE 105

Conclusion

slide-106
SLIDE 106

Summary

  • Tree space makes for some interesting problems that takes

away some standard statistical tricks.

  • Model selection typically relies on multiple metrics
  • Bootstrapping is a slow, biased but conservative way to estimate

the support for a given branch in your tree.

  • Likelihood Testing is powerful but must be used with care.
  • Comparing trees directly is non-trivial due to tree-space.

37

slide-107
SLIDE 107

Summary

  • Tree space makes for some interesting problems that takes

away some standard statistical tricks.

  • Model selection typically relies on multiple metrics
  • Bootstrapping is a slow, biased but conservative way to estimate

the support for a given branch in your tree.

  • Likelihood Testing is powerful but must be used with care.
  • Comparing trees directly is non-trivial due to tree-space.

37

slide-108
SLIDE 108

Summary

  • Tree space makes for some interesting problems that takes

away some standard statistical tricks.

  • Model selection typically relies on multiple metrics
  • Bootstrapping is a slow, biased but conservative way to estimate

the support for a given branch in your tree.

  • Likelihood Testing is powerful but must be used with care.
  • Comparing trees directly is non-trivial due to tree-space.

37

slide-109
SLIDE 109

Summary

  • Tree space makes for some interesting problems that takes

away some standard statistical tricks.

  • Model selection typically relies on multiple metrics
  • Bootstrapping is a slow, biased but conservative way to estimate

the support for a given branch in your tree.

  • Likelihood Testing is powerful but must be used with care.
  • Comparing trees directly is non-trivial due to tree-space.

37

slide-110
SLIDE 110

Summary

  • Tree space makes for some interesting problems that takes

away some standard statistical tricks.

  • Model selection typically relies on multiple metrics
  • Bootstrapping is a slow, biased but conservative way to estimate

the support for a given branch in your tree.

  • Likelihood Testing is powerful but must be used with care.
  • Comparing trees directly is non-trivial due to tree-space.

37

slide-111
SLIDE 111

Questions?

37

slide-112
SLIDE 112

References i

Leonard, G. (2010). Development of fusion and duplication finder blast (fdfblast): a systematic tool to detect differentially distributed gene fusions and resolve trifurcations in the tree of life. Marwick, B. (2012). A cladistic evaluation of ancient thai bronze buddha images: six tests for a phylogenetic signal in the griswold collection. Connecting empires, pages 159–176. Nickle, D. C., Heath, L., Jensen, M. A., Gilbert, P. B., Mullins, J. I., and Pond, S. L. K. (2007). Hiv-specific probabilistic models of protein evolution. PLoS One, 2(6):e503.

38

slide-113
SLIDE 113

References ii

Richards, T. A., Soanes, D. M., Foster, P. G., Leonard, G., Thornton,

  • C. R., and Talbot, N. J. (2009).

Phylogenomic analysis demonstrates a pattern of rare and ancient horizontal gene transfer between plants and fungi. The Plant Cell, 21(7):1897–1911.

39