Regression Noise model and likelihood Given a dataset D = { x n , y - - PowerPoint PPT Presentation

regression
SMART_READER_LITE
LIVE PREVIEW

Regression Noise model and likelihood Given a dataset D = { x n , y - - PowerPoint PPT Presentation

Linear Regression Regression Noise model and likelihood Given a dataset D = { x n , y n } S n =1 , where x n = { x n, 1 , . . . , x n,S } is S dimensional, fit parameters of a regressor f with added Gaussian noise: 0 , 2


slide-1
SLIDE 1

Linear Regression

Regression

Noise model and likelihood

◮ Given a dataset D = {xn, yn}S n=1, where xn = {xn,1, . . . , xn,S} is S

dimensional, fit parameters θ of a regressor f with added Gaussian noise: yn = f(xn; θ) + ǫn where p(ǫ | σ2) = N

  • ǫ
  • 0, σ2

.

◮ Equivalent likelihood formulation:

p(y | X) =

N

  • n=1

N

  • yn
  • f(xn), σ2

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 1

slide-2
SLIDE 2

Linear Regression

Regression

Choosing a regressor

◮ Choose f to be linear:

p(y | X) =

N

  • n=1

N

  • yn
  • xn · θ + c, σ2

◮ Consider bias free case, c = 0,

  • therwise include an additional

column of ones in each xn.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 2

slide-3
SLIDE 3

Linear Regression

Regression

Choosing a regressor

◮ Choose f to be linear:

p(y | X) =

N

  • n=1

N

  • yn
  • xn · θ + c, σ2

◮ Consider bias free case, c = 0,

  • therwise include an additional

column of ones in each xn.

Equivalent graphical model Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 2

slide-4
SLIDE 4

Linear Regression

Linear Regression

Maximum likelihood

◮ Taking the logarithm, we obtain

ln p(y | θσ2) =

N

  • n=1

ln N

  • yn
  • xn · θ, σ2

= −N 2 ln 2πσ2 − 1 2σ2

N

  • n=1

(yn − xn · θ)2

  • Sum of squares

◮ The likelihood is maximized when the squared error is minimized. ◮ Least squares and maximum likelihood are equivalent.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 3

slide-5
SLIDE 5

Linear Regression

Linear Regression

Maximum likelihood

◮ Taking the logarithm, we obtain

ln p(y | θσ2) =

N

  • n=1

ln N

  • yn
  • xn · θ, σ2

= −N 2 ln 2πσ2 − 1 2σ2

N

  • n=1

(yn − xn · θ)2

  • Sum of squares

◮ The likelihood is maximized when the squared error is minimized. ◮ Least squares and maximum likelihood are equivalent.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 3

slide-6
SLIDE 6

Linear Regression

Linear Regression

Maximum likelihood

◮ Taking the logarithm, we obtain

ln p(y | θσ2) =

N

  • n=1

ln N

  • yn
  • xn · θ, σ2

= −N 2 ln 2πσ2 − 1 2σ2

N

  • n=1

(yn − xn · θ)2

  • Sum of squares

◮ The likelihood is maximized when the squared error is minimized. ◮ Least squares and maximum likelihood are equivalent.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 3

slide-7
SLIDE 7

Linear Regression

Linear Regression and Least Squares

y x f(xn, w) y

n

xn

(C.M. Bishop, Pattern Recognition and Machine Learning)

E(θ) = 1 2

N

  • n=1

(yn − xn · θ)2

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 4

slide-8
SLIDE 8

Linear Regression

Linear Regression and Least Squares

◮ Derivative w.r.t a single weight entry θi

d dθi ln p(y | θ, σ2) = d dθi

  • − 1

2σ2

N

  • n=1

(yn − xn · θ)2

  • ◮ Set gradient w.r.t. θ to zero

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 5

slide-9
SLIDE 9

Linear Regression

Linear Regression and Least Squares

◮ Derivative w.r.t a single weight entry θi

d dθi ln p(y | θ, σ2) = d dθi

  • − 1

2σ2

N

  • n=1

(yn − xn · θ)2

  • ◮ Set gradient w.r.t. θ to zero

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 5

slide-10
SLIDE 10

Linear Regression

Linear Regression and Least Squares

◮ Derivative w.r.t a single weight entry θi

d dθi ln p(y | θ, σ2) = d dθi

  • − 1

2σ2

N

  • n=1

(yn − xn · θ)2

  • ◮ Set gradient w.r.t. θ to zero

∇θ ln p(y | θ, σ2) = 1 σ2

N

  • n=1

(yn − xn · θ)xT

n = 0

= ⇒ θML =?

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 5

slide-11
SLIDE 11

Linear Regression

Linear Regression and Least Squares

◮ Derivative w.r.t. a single weight entry θi

d dθi ln p(y | θ, σ2) = d dθi

  • − 1

2σ2

N

  • n=1

(yn − xn · θ)2

  • = 1

σ2

N

  • n=1

(yn − xn · θ)xi

◮ Set gradient w.r.t. θ to zero

∇θ ln p(y | θ, σ2) = 1 σ2

N

  • n=1

(yn − xn · θ)xT

n = 0

= ⇒ θML = (XTX)−1XT

  • Pseudo inverse

y

◮ Here, the matrix X is defined as X =

  x1,1 . . . x1, S . . . . . . . . . xN,1 . . . xN,S  

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 6

slide-12
SLIDE 12

Linear Regression

Linear Regression and Least Squares

◮ Derivative w.r.t. a single weight entry θi

d dθi ln p(y | θ, σ2) = d dθi

  • − 1

2σ2

N

  • n=1

(yn − xn · θ)2

  • = 1

σ2

N

  • n=1

(yn − xn · θ)xi

◮ Set gradient w.r.t. θ to zero

∇θ ln p(y | θ, σ2) = 1 σ2

N

  • n=1

(yn − xn · θ)xT

n = 0

= ⇒ θML = (XTX)−1XT

  • Pseudo inverse

y

◮ Here, the matrix X is defined as X =

  x1,1 . . . x1, S . . . . . . . . . xN,1 . . . xN,S  

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 6

slide-13
SLIDE 13

Linear Regression

Linear Regression and Least Squares

◮ Derivative w.r.t. a single weight entry θi

d dθi ln p(y | θ, σ2) = d dθi

  • − 1

2σ2

N

  • n=1

(yn − xn · θ)2

  • = 1

σ2

N

  • n=1

(yn − xn · θ)xi

◮ Set gradient w.r.t. θ to zero

∇θ ln p(y | θ, σ2) = 1 σ2

N

  • n=1

(yn − xn · θ)xT

n = 0

= ⇒ θML = (XTX)−1XT

  • Pseudo inverse

y

◮ Here, the matrix X is defined as X =

  x1,1 . . . x1, S . . . . . . . . . xN,1 . . . xN,S  

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 6

slide-14
SLIDE 14

Hypothesis Testing

Hypothesis Testing

Example:

◮ Given a sample

D = {x1, . . . , xN}.

◮ Test whether H0 : θs = 0 (null

hypothesis) or H1 : θs = 0 (alternative hypothesis) is true.

◮ To show that θs = 0 we can

perform a statistical test that tries to reject H0.

◮ type 1 error: H0 is rejected but

does hold.

◮ type 2 error: H0 is accepted

but does not hold.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 7

slide-15
SLIDE 15

Hypothesis Testing

Hypothesis Testing

Example:

◮ Given a sample

D = {x1, . . . , xN}.

◮ Test whether H0 : θs = 0 (null

hypothesis) or H1 : θs = 0 (alternative hypothesis) is true.

◮ To show that θs = 0 we can

perform a statistical test that tries to reject H0.

◮ type 1 error: H0 is rejected but

does hold.

◮ type 2 error: H0 is accepted

but does not hold.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 7

slide-16
SLIDE 16

Hypothesis Testing

Hypothesis Testing

Example:

◮ Given a sample

D = {x1, . . . , xN}.

◮ Test whether H0 : θs = 0 (null

hypothesis) or H1 : θs = 0 (alternative hypothesis) is true.

◮ To show that θs = 0 we can

perform a statistical test that tries to reject H0.

◮ type 1 error: H0 is rejected but

does hold.

◮ type 2 error: H0 is accepted

but does not hold.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 7

slide-17
SLIDE 17

Hypothesis Testing

Hypothesis Testing

Example:

◮ Given a sample

D = {x1, . . . , xN}.

◮ Test whether H0 : θs = 0 (null

hypothesis) or H1 : θs = 0 (alternative hypothesis) is true.

◮ To show that θs = 0 we can

perform a statistical test that tries to reject H0.

◮ type 1 error: H0 is rejected but

does hold.

◮ type 2 error: H0 is accepted

but does not hold.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 7

slide-18
SLIDE 18

Hypothesis Testing

Hypothesis Testing

Example:

◮ Given a sample

D = {x1, . . . , xN}.

◮ Test whether H0 : θs = 0 (null

hypothesis) or H1 : θs = 0 (alternative hypothesis) is true.

◮ To show that θs = 0 we can

perform a statistical test that tries to reject H0.

◮ type 1 error: H0 is rejected but

does hold.

◮ type 2 error: H0 is accepted

but does not hold.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 7

slide-19
SLIDE 19

Hypothesis Testing

Hypothesis Testing

◮ Given a sample

D = {x1, . . . , xN}.

◮ Test whether H0 : θs = 0 (null

hypothesis) or H1 : θs = 0 (alternative hypothesis) is true.

◮ The significance level α defines

the threshold and the sensitivity

  • f the test. This equals the

probability of a type-1 error.

◮ Usually decision is based on a

test statistic.

◮ The critical region defines the

values of the test statistic that lead to a rejection of the test.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 8

slide-20
SLIDE 20

Hypothesis Testing

Hypothesis Testing

◮ Given a sample

D = {x1, . . . , xN}.

◮ Test whether H0 : θs = 0 (null

hypothesis) or H1 : θs = 0 (alternative hypothesis) is true.

◮ The significance level α defines

the threshold and the sensitivity

  • f the test. This equals the

probability of a type-1 error.

◮ Usually decision is based on a

test statistic.

◮ The critical region defines the

values of the test statistic that lead to a rejection of the test.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 8

slide-21
SLIDE 21

Hypothesis Testing

Hypothesis Testing

◮ Given a sample

D = {x1, . . . , xN}.

◮ Test whether H0 : θs = 0 (null

hypothesis) or H1 : θs = 0 (alternative hypothesis) is true.

◮ The significance level α defines

the threshold and the sensitivity

  • f the test. This equals the

probability of a type-1 error.

◮ Usually decision is based on a

test statistic.

◮ The critical region defines the

values of the test statistic that lead to a rejection of the test.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 8

slide-22
SLIDE 22

Hypothesis Testing

Hypothesis Testing

◮ Given a sample

D = {x1, . . . , xN}.

◮ Test whether H0 : θs = 0 (null

hypothesis) or H1 : θs = 0 (alternative hypothesis) is true.

◮ The significance level α defines

the threshold and the sensitivity

  • f the test. This equals the

probability of a type-1 error.

◮ Usually decision is based on a

test statistic.

◮ The critical region defines the

values of the test statistic that lead to a rejection of the test.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 8

slide-23
SLIDE 23

Hypothesis Testing

Testing in Linear Regression

Likelihood Ratio Test

p(y | X) =

N

  • n=1

N

  • yn
  • xn · θ + xn,sβ, σ2

◮ xn,s: SNP to be tested ◮ xn: regression covariates (including

bias term)

◮ Race ◮ Known background SNPs ◮ Environment Equivalent graphical model

xn: regression covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 9

slide-24
SLIDE 24

Hypothesis Testing

Testing in Linear Regression

Likelihood Ratio Test

p(y | X) =

N

  • n=1

N

  • yn
  • xn · θ + xn,sβ, σ2

◮ xn,s: SNP to be tested ◮ xn: regression covariates (including

bias term)

◮ Race ◮ Known background SNPs ◮ Environment Equivalent graphical model

xn: regression covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 9

slide-25
SLIDE 25

Hypothesis Testing

Testing in Linear Regression

Likelihood Ratio Test

p(y | X) =

N

  • n=1

N

  • yn
  • xn · θ + xn,sβ, σ2

◮ xn,s: SNP to be tested ◮ xn: regression covariates (including

bias term)

◮ Race ◮ Known background SNPs ◮ Environment Equivalent graphical model

xn: regression covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 9

slide-26
SLIDE 26

Hypothesis Testing

Testing in Linear Regression

Likelihood Ratio Test

p(y | X) =

N

  • n=1

N

  • yn
  • xn · θ + xn,sβ, σ2

◮ Test H0 : β = 0 ◮ The ratio of the ML estimator and

the ML0 estimator restricted to H0 is a common test statistic.

Equivalent graphical model

xn: regression covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 10

slide-27
SLIDE 27

Hypothesis Testing

Testing in Linear Regression

Likelihood Ratio Test

p(y | X) =

N

  • n=1

N

  • yn
  • xn · θ + xn,sβ, σ2

◮ Test H0 : β = 0 ◮ The ratio of the ML estimator and

the ML0 estimator restricted to H0 is a common test statistic.

Equivalent graphical model

xn: regression covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 10

slide-28
SLIDE 28

Hypothesis Testing

Testing in Linear Regression

Likelihood Ratio Test

p(y | X) =

N

  • n=1

N

  • yn
  • xn · θ + xn,sβ, σ2

◮ Test H0 : β = 0 ◮ The ratio of the ML estimator and

the ML0 estimator restricted to H0 is a common test statistic. N

n=1 N

  • yn
  • xn · θML + xn,sβML, σ2

ML

  • N

n=1 N

  • yn
  • xn · θML0 + xn,s0, σ2

ML0

  • Equivalent graphical model

xn: regression covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 10

slide-29
SLIDE 29

Hypothesis Testing

P-value

definition

◮ Probability of observing a test statistic at least as extreme (e.g.

likelihood ratio statistic), given that H0 is true.

◮ Significance level α becomes threshold on P-value. ◮ Need to know the null distribution of test statistics. (usually

unknown)

◮ Possible to generate artificial null-distribution by permutations

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 11

slide-30
SLIDE 30

Hypothesis Testing

P-value

definition

◮ Probability of observing a test statistic at least as extreme (e.g.

likelihood ratio statistic), given that H0 is true.

◮ Significance level α becomes threshold on P-value. ◮ Need to know the null distribution of test statistics. (usually

unknown)

◮ Possible to generate artificial null-distribution by permutations

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 11

slide-31
SLIDE 31

Hypothesis Testing

P-value

definition

◮ Probability of observing a test statistic at least as extreme (e.g.

likelihood ratio statistic), given that H0 is true.

◮ Significance level α becomes threshold on P-value. ◮ Need to know the null distribution of test statistics. (usually

unknown)

◮ Possible to generate artificial null-distribution by permutations

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 11

slide-32
SLIDE 32

Hypothesis Testing

P-value

definition

◮ Probability of observing a test statistic at least as extreme (e.g.

likelihood ratio statistic), given that H0 is true.

◮ Significance level α becomes threshold on P-value. ◮ Need to know the null distribution of test statistics. (usually

unknown)

◮ Possible to generate artificial null-distribution by permutations

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 11

slide-33
SLIDE 33

Hypothesis Testing

P-value

Permutation procedure

Repeat M times:

◮ Permute phenotype y and

covariates x jointly over individuals.

◮ Compute permuted test statistic ◮ Add test statistic to emprirical null

distribution

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 12

slide-34
SLIDE 34

Hypothesis Testing

P-value

Permutation procedure

Repeat M times:

◮ Permute phenotype y and

covariates x jointly over individuals.

◮ Compute permuted test statistic ◮ Add test statistic to emprirical null

distribution

covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 12

slide-35
SLIDE 35

Hypothesis Testing

P-value

Permutation procedure

Repeat M times:

◮ Permute phenotype y and

covariates x jointly over individuals.

◮ Compute permuted test statistic ◮ Add test statistic to emprirical null

distribution

covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 12

slide-36
SLIDE 36

Hypothesis Testing

P-value

Permutation procedure

Repeat M times:

◮ Permute phenotype y and

covariates x jointly over individuals.

◮ Compute permuted test statistic ◮ Add test statistic to emprirical null

distribution

covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 12

slide-37
SLIDE 37

Hypothesis Testing

P-value

Permutation procedure

Repeat M times:

◮ Permute phenotype y and

covariates x jointly over individuals.

◮ Compute permuted test statistic ◮ Add test statistic to emprirical null

distribution The P-value is the quantile of real test statistic in artificial null distribution.

covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 12

slide-38
SLIDE 38

Hypothesis Testing

Testing in Linear Regression

Likelihood Ratio Test revisited

◮ Can equivalently compute

log-likelihood ratio:

◮ Wilks’ theorem: 2LR follows a

Chi-square distribution with 1 degree of freedom.

◮ P-value = 1-CDF(2LR).

Equivalent graphical model

xn: regression covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 13

slide-39
SLIDE 39

Hypothesis Testing

Testing in Linear Regression

Likelihood Ratio Test revisited

◮ Can equivalently compute

log-likelihood ratio:

LR =

N

  • n=1

log N

  • yn
  • xn · θML + xn,sβML, σ2

ML

N

  • n=1

log N

  • yn
  • xn · θML0, σ2

ML0

  • ◮ Wilks’ theorem: 2LR follows a

Chi-square distribution with 1 degree of freedom.

◮ P-value = 1-CDF(2LR).

Equivalent graphical model

xn: regression covariates

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 13

slide-40
SLIDE 40

Hypothesis Testing

Testing in Linear Regression

Likelihood Ratio Test revisited

◮ Can equivalently compute

log-likelihood ratio:

LR =

N

  • n=1

log N

  • yn
  • xn · θML + xn,sβML, σ2

ML

N

  • n=1

log N

  • yn
  • xn · θML0, σ2

ML0

  • ◮ Wilks’ theorem: 2LR follows a

Chi-square distribution with 1 degree of freedom.

◮ P-value = 1-CDF(2LR).

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 13

slide-41
SLIDE 41

Hypothesis Testing

Testing in Linear Regression

Likelihood Ratio Test revisited

◮ Can equivalently compute

log-likelihood ratio:

LR =

N

  • n=1

log N

  • yn
  • xn · θML + xn,sβML, σ2

ML

N

  • n=1

log N

  • yn
  • xn · θML0, σ2

ML0

  • ◮ Wilks’ theorem: 2LR follows a

Chi-square distribution with 1 degree of freedom.

◮ P-value = 1-CDF(2LR).

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 13

slide-42
SLIDE 42

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Motivation

◮ Significance level α equals

probability of type-1 error.

◮ In GWAS we perform S = 106 tests ◮ At α = 0.01 we would expect 10000

type-1 errors!

◮ Probability of at least 1 type-1 error

is 1 − (1 − α)S → 1.

◮ Individual P-values < 0.01 are not

significant anymore.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 14

slide-43
SLIDE 43

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Motivation

◮ Significance level α equals

probability of type-1 error.

◮ In GWAS we perform S = 106 tests ◮ At α = 0.01 we would expect 10000

type-1 errors!

◮ Probability of at least 1 type-1 error

is 1 − (1 − α)S → 1.

◮ Individual P-values < 0.01 are not

significant anymore.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 14

slide-44
SLIDE 44

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Motivation

◮ Significance level α equals

probability of type-1 error.

◮ In GWAS we perform S = 106 tests ◮ At α = 0.01 we would expect 10000

type-1 errors!

◮ Probability of at least 1 type-1 error

is 1 − (1 − α)S → 1.

◮ Individual P-values < 0.01 are not

significant anymore.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 14

slide-45
SLIDE 45

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Motivation

◮ Significance level α equals

probability of type-1 error.

◮ In GWAS we perform S = 106 tests ◮ At α = 0.01 we would expect 10000

type-1 errors!

◮ Probability of at least 1 type-1 error

is 1 − (1 − α)S → 1.

◮ Individual P-values < 0.01 are not

significant anymore.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 14

slide-46
SLIDE 46

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Motivation

◮ Significance level α equals

probability of type-1 error.

◮ In GWAS we perform S = 106 tests ◮ At α = 0.01 we would expect 10000

type-1 errors!

◮ Probability of at least 1 type-1 error

is 1 − (1 − α)S → 1.

◮ Individual P-values < 0.01 are not

significant anymore.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 14

slide-47
SLIDE 47

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Motivation

◮ Significance level α equals

probability of type-1 error.

◮ In GWAS we perform S = 106 tests ◮ At α = 0.01 we would expect 10000

type-1 errors!

◮ Probability of at least 1 type-1 error

is 1 − (1 − α)S → 1.

◮ Individual P-values < 0.01 are not

significant anymore. Need to correct for multiple hypothesis testing!

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 14

slide-48
SLIDE 48

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Family-Wise Error Rate (FWER)

◮ Probability of at least one type-2

error.

◮ Correct by bounding the FWER. ◮ Bonferroni correction: PB = P ∗ S ◮ Equivalently P < α

S significant.

◮ Bounds the FWER 1 − (1 − α/S)S

by α

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 15

slide-49
SLIDE 49

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Family-Wise Error Rate (FWER)

◮ Probability of at least one type-2

error.

◮ Correct by bounding the FWER. ◮ Bonferroni correction: PB = P ∗ S ◮ Equivalently P < α

S significant.

◮ Bounds the FWER 1 − (1 − α/S)S

by α

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 15

slide-50
SLIDE 50

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Family-Wise Error Rate (FWER)

◮ Probability of at least one type-2

error.

◮ Correct by bounding the FWER. ◮ Bonferroni correction: PB = P ∗ S ◮ Equivalently P < α

S significant.

◮ Bounds the FWER 1 − (1 − α/S)S

by α

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 15

slide-51
SLIDE 51

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Family-Wise Error Rate (FWER)

◮ Probability of at least one type-2

error.

◮ Correct by bounding the FWER. ◮ Bonferroni correction: PB = P ∗ S ◮ Equivalently P < α

S significant.

◮ Bounds the FWER 1 − (1 − α/S)S

by α

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 15

slide-52
SLIDE 52

Multiple Hypothesis Testing

Multiple Hypothesis Testing

Family-Wise Error Rate (FWER)

◮ Probability of at least one type-2

error.

◮ Correct by bounding the FWER. ◮ Bonferroni correction: PB = P ∗ S ◮ Equivalently P < α

S significant.

◮ Bounds the FWER 1 − (1 − α/S)S

by α

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 15

slide-53
SLIDE 53

Multiple Hypothesis Testing

False Discovery Rate (FDR)

◮ FWER based correction

(Bonferroni) leads to very conservative significance thresholds.

◮ Because of the abundance of tests

we might be willing to accept a few false positives.

◮ Intuitive definition of the FDR:

◮ E

  • FP

FP + TP

  • ◮ But: this can not be bounded when

H0 always true (FN + TP = 0). In this case E

  • FP

FP + TP

  • = 1.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 16

slide-54
SLIDE 54

Multiple Hypothesis Testing

False Discovery Rate (FDR)

◮ FWER based correction

(Bonferroni) leads to very conservative significance thresholds.

◮ Because of the abundance of tests

we might be willing to accept a few false positives.

◮ Intuitive definition of the FDR:

◮ E

  • FP

FP + TP

  • ◮ But: this can not be bounded when

H0 always true (FN + TP = 0). In this case E

  • FP

FP + TP

  • = 1.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 16

slide-55
SLIDE 55

Multiple Hypothesis Testing

False Discovery Rate (FDR)

◮ FWER based correction

(Bonferroni) leads to very conservative significance thresholds.

◮ Because of the abundance of tests

we might be willing to accept a few false positives.

◮ Intuitive definition of the FDR:

◮ E

  • FP

FP + TP

  • ◮ But: this can not be bounded when

H0 always true (FN + TP = 0). In this case E

  • FP

FP + TP

  • = 1.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 16

slide-56
SLIDE 56

Multiple Hypothesis Testing

False Discovery Rate (FDR)

◮ FWER based correction

(Bonferroni) leads to very conservative significance thresholds.

◮ Because of the abundance of tests

we might be willing to accept a few false positives.

◮ Intuitive definition of the FDR:

◮ E

  • FP

FP + TP

  • ◮ But: this can not be bounded when

H0 always true (FN + TP = 0). In this case E

  • FP

FP + TP

  • = 1.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 16

slide-57
SLIDE 57

Multiple Hypothesis Testing

False Discovery Rate (FDR)

◮ FWER based correction

(Bonferroni) leads to very conservative significance thresholds.

◮ Because of the abundance of tests

we might be willing to accept a few false positives.

◮ Intuitive definition of the FDR:

◮ E

  • FP

FP + TP

  • ◮ But: this can not be bounded when

H0 always true (FN + TP = 0). In this case E

  • FP

FP + TP

  • = 1.

H0 holds H0 doesn’t hold H0 accepted true negatives false negatives type-2 error H0 rejected false positives true positives type-1 error Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 16

slide-58
SLIDE 58

Multiple Hypothesis Testing

False Discovery Rate (FDR)

◮ If (FP + TP) = 0, then

FDR =E

  • FP

FP + TP |(FP + TP) > 0

  • ∗ P(FP + TP > 0)

=E [1|(FP + TP) > 0] ∗ P(FP + TP > 0) + 0 ∗ P(FP + TP = 0) =E [Q] with Q = FP FP + TP if (FP + TP) > 0 and 0, otherwise.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 17

slide-59
SLIDE 59

Multiple Hypothesis Testing

False Discovery Rate (FDR)

◮ Actual definition of the FDR:

FDR =E

  • FP

FP + TP |(FP + TP) > 0

  • ∗ P(FP + TP > 0)

=E [1|(FP + TP) > 0] ∗ P(FP + TP > 0) =P(FP + TP > 0) = FWER

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 18

slide-60
SLIDE 60

Multiple Hypothesis Testing

False Discovery Rate (FDR)

Test Procedure (Benjamini & Hochberg, 1995)

Input:

◮ P-values: P1, . . . , PS (need to be

independent)

◮ FDR thresold q

  • Christoph Lippert

GWAS II: Linear models and Statistics Summer 2011 19

slide-61
SLIDE 61

Multiple Hypothesis Testing

False Discovery Rate (FDR)

Test Procedure (Benjamini & Hochberg, 1995)

Input:

◮ P-values: P1, . . . , PS (need to be

independent)

◮ FDR thresold q

Algorithm:

◮ Sort: P(1) ≤ P(2) ≤ · · · ≤ P(S) ◮ k = argmax i

P(i) ≤ i S q = αS

◮ Reject all Ps with Ps < αS

1 p(i) i

" !

1 3 5 7 10

$

  • Ablehnungsbereiche für...

$

  • Christoph Lippert

GWAS II: Linear models and Statistics Summer 2011 19

slide-62
SLIDE 62

Multiple Hypothesis Testing

False Discovery Rate (FDR)

Test Procedure (Benjamini & Hochberg, 1995)

Input:

◮ P-values: P1, . . . , PS (need to be

independent)

◮ FDR thresold q

Algorithm:

◮ Sort: P(1) ≤ P(2) ≤ · · · ≤ P(S) ◮ k = argmax i

P(i) ≤ i S q = αS

◮ Reject all Ps with Ps < αS ◮ For this procedure holds:

FDR ≤ FP + TN S q ≤ q

1 p(i) i

" !

1 3 5 7 10

$

  • Ablehnungsbereiche für...

$

  • Christoph Lippert

GWAS II: Linear models and Statistics Summer 2011 19

slide-63
SLIDE 63

Multiple Hypothesis Testing

Model Checking

◮ Do my estimated P-values match

the true null distribution?

◮ By definition uniformly distributed

under null distribution.

◮ Do the empirical results match my

assumptions on the null model?

◮ In GWAS we perform a large

number of tests. (usually in the

  • rder of 106)

◮ Use the strong prior knowledge that

in GWAS almost all of the test SNPs have no effect on the phenotype.

◮ Empirical test statistics should

follow the null distribution

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 20

slide-64
SLIDE 64

Multiple Hypothesis Testing

Model Checking

◮ Do my estimated P-values match

the true null distribution?

◮ By definition uniformly distributed

under null distribution.

◮ Do the empirical results match my

assumptions on the null model?

◮ In GWAS we perform a large

number of tests. (usually in the

  • rder of 106)

◮ Use the strong prior knowledge that

in GWAS almost all of the test SNPs have no effect on the phenotype.

◮ Empirical test statistics should

follow the null distribution

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 20

slide-65
SLIDE 65

Multiple Hypothesis Testing

Model Checking

◮ Do my estimated P-values match

the true null distribution?

◮ By definition uniformly distributed

under null distribution.

◮ Do the empirical results match my

assumptions on the null model?

◮ In GWAS we perform a large

number of tests. (usually in the

  • rder of 106)

◮ Use the strong prior knowledge that

in GWAS almost all of the test SNPs have no effect on the phenotype.

◮ Empirical test statistics should

follow the null distribution

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 20

slide-66
SLIDE 66

Multiple Hypothesis Testing

Model Checking

◮ Do my estimated P-values match

the true null distribution?

◮ By definition uniformly distributed

under null distribution.

◮ Do the empirical results match my

assumptions on the null model?

◮ In GWAS we perform a large

number of tests. (usually in the

  • rder of 106)

◮ Use the strong prior knowledge that

in GWAS almost all of the test SNPs have no effect on the phenotype.

◮ Empirical test statistics should

follow the null distribution

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 20

slide-67
SLIDE 67

Multiple Hypothesis Testing

Model Checking

◮ Do my estimated P-values match

the true null distribution?

◮ By definition uniformly distributed

under null distribution.

◮ Do the empirical results match my

assumptions on the null model?

◮ In GWAS we perform a large

number of tests. (usually in the

  • rder of 106)

◮ Use the strong prior knowledge that

in GWAS almost all of the test SNPs have no effect on the phenotype.

◮ Empirical test statistics should

follow the null distribution

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 20

slide-68
SLIDE 68

Multiple Hypothesis Testing

Model Checking

QQ-plot

Compare quantiles of the empirical test statistic distribution to assumed null distribution.

◮ Sort test statistics ◮ Plot test statisitcs against (y-axis)

quantiles of the theoretical null-distribution (x-axis)

◮ for example: 2LR vs. χ2

1

◮ If the plot is close to the diagonal,

the distributions match up

◮ Deviation from the diagonal

indicates inflation or deflation of test statistics.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 21

slide-69
SLIDE 69

Multiple Hypothesis Testing

Model Checking

QQ-plot

Compare quantiles of the empirical test statistic distribution to assumed null distribution.

◮ Sort test statistics ◮ Plot test statisitcs against (y-axis)

quantiles of the theoretical null-distribution (x-axis)

◮ for example: 2LR vs. χ2

1

◮ If the plot is close to the diagonal,

the distributions match up

◮ Deviation from the diagonal

indicates inflation or deflation of test statistics.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 21

slide-70
SLIDE 70

Multiple Hypothesis Testing

Model Checking

QQ-plot

Compare quantiles of the empirical test statistic distribution to assumed null distribution.

◮ Sort test statistics ◮ Plot test statisitcs against (y-axis)

quantiles of the theoretical null-distribution (x-axis)

◮ for example: 2LR vs. χ2

1

◮ If the plot is close to the diagonal,

the distributions match up

◮ Deviation from the diagonal

indicates inflation or deflation of test statistics.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 21

slide-71
SLIDE 71

Multiple Hypothesis Testing

Model Checking

QQ-plot

Compare quantiles of the empirical test statistic distribution to assumed null distribution.

◮ Sort test statistics ◮ Plot test statisitcs against (y-axis)

quantiles of the theoretical null-distribution (x-axis)

◮ for example: 2LR vs. χ2

1

◮ If the plot is close to the diagonal,

the distributions match up

◮ Deviation from the diagonal

indicates inflation or deflation of test statistics.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 21

slide-72
SLIDE 72

Multiple Hypothesis Testing

Model Checking

QQ-plot

Compare quantiles of the empirical test statistic distribution to assumed null distribution.

◮ Sort test statistics ◮ Plot test statisitcs against (y-axis)

quantiles of the theoretical null-distribution (x-axis)

◮ for example: 2LR vs. χ2

1

◮ If the plot is close to the diagonal,

the distributions match up

◮ Deviation from the diagonal

indicates inflation or deflation of test statistics.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 21

slide-73
SLIDE 73

Multiple Hypothesis Testing

Correction for inflation

Genomic control (λGC)

◮ Ratio of the 50% quantiles between

theoretical distribution and test-statistics known as the genomic inflation factor λGC.

◮ λGC should be close to 1. ◮ Estimate degree of inflation

(deflation) from this ratio.

◮ Adjust for degree of inflation by

dividing all statistics by ratio of the median (50%-quantile).

◮ This procedure yields conservative

estimates of the P-value distribution null-distribution.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 22

slide-74
SLIDE 74

Multiple Hypothesis Testing

Correction for inflation

Genomic control (λGC)

◮ Ratio of the 50% quantiles between

theoretical distribution and test-statistics known as the genomic inflation factor λGC.

◮ λGC should be close to 1. ◮ Estimate degree of inflation

(deflation) from this ratio.

◮ Adjust for degree of inflation by

dividing all statistics by ratio of the median (50%-quantile).

◮ This procedure yields conservative

estimates of the P-value distribution null-distribution.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 22

slide-75
SLIDE 75

Multiple Hypothesis Testing

Correction for inflation

Genomic control (λGC)

◮ Ratio of the 50% quantiles between

theoretical distribution and test-statistics known as the genomic inflation factor λGC.

◮ λGC should be close to 1. ◮ Estimate degree of inflation

(deflation) from this ratio.

◮ Adjust for degree of inflation by

dividing all statistics by ratio of the median (50%-quantile).

◮ This procedure yields conservative

estimates of the P-value distribution null-distribution.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 22

slide-76
SLIDE 76

Multiple Hypothesis Testing

Correction for inflation

Genomic control (λGC)

◮ Ratio of the 50% quantiles between

theoretical distribution and test-statistics known as the genomic inflation factor λGC.

◮ λGC should be close to 1. ◮ Estimate degree of inflation

(deflation) from this ratio.

◮ Adjust for degree of inflation by

dividing all statistics by ratio of the median (50%-quantile).

◮ This procedure yields conservative

estimates of the P-value distribution null-distribution.

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 22

slide-77
SLIDE 77

Multiple Hypothesis Testing

Correction for inflation

Genomic control (λGC)

◮ Ratio of the 50% quantiles between

theoretical distribution and test-statistics known as the genomic inflation factor λGC.

◮ λGC should be close to 1. ◮ Estimate degree of inflation

(deflation) from this ratio.

◮ Adjust for degree of inflation by

dividing all statistics by ratio of the median (50%-quantile).

◮ This procedure yields conservative

estimates of the P-value distribution null-distribution.

Does not make P-values uniform!

Christoph Lippert GWAS II: Linear models and Statistics Summer 2011 22