Leveraging prior information and group structure for false discovery - - PowerPoint PPT Presentation

leveraging prior information and group structure for
SMART_READER_LITE
LIVE PREVIEW

Leveraging prior information and group structure for false discovery - - PowerPoint PPT Presentation

Leveraging prior information and group structure for false discovery rate control Rina Foygel Barber Dept. of Statistics, University of Chicago http://www.stat.uchicago.edu/~rina/ Multiple comparisons & FDR control When testing n different


slide-1
SLIDE 1

Leveraging prior information and group structure for false discovery rate control

Rina Foygel Barber

  • Dept. of Statistics, University of Chicago

http://www.stat.uchicago.edu/~rina/

slide-2
SLIDE 2

Multiple comparisons & FDR control

When testing n different questions simultaneously, how to determine which effects are significant?

  • False discovery proportion:

FDP = # false discoveries total # discoveries = |H0 ∩ S| | S|

  • False discovery rate:

FDR = E [FDP]

2/29

slide-3
SLIDE 3

Multiple comparisons & FDR control

Benjamini-Hochberg (BH) procedure (1995): set a data-dependent threshold for rejecting p-values, to adapt to the amount of signal present in the data

  • If we reject all p-values below a fixed threshold t,

FDP(t) ≈ t · |H0| #{i : Pi ≤ t} = FDP(t)

  • Choose adaptive threshold: max t with

FDP(t) ≤ α

  • Guaranteed to control FDR at level α

if p-values are independent or positively dependent (PRDS)

Benjamini & Hochberg 1995; Benjamini & Yekutieli 2001 3/29

slide-4
SLIDE 4

Multiple comparisons & FDR control

How can we incorporate additional information into the FDR control problem?

  • If some of the hypotheses are more likely to contain true signals,

should we give them priority?

  • If the hypotheses have a grouped / clustered / hierarchical structure,

how can we take this into account?

4/29

slide-5
SLIDE 5

Outline

  • 1. Accumulation tests: testing a ranked list of hypotheses
  • Joint work with Ang Li
  • 2. The p-filter: FDR control across groups
  • Joint work with Aaditya Ramdas

5/29

slide-6
SLIDE 6

Ordered hypothesis testing

Setting: a multiple comparisons problem with a pre-defined ordering. p-values: P1, P2, P3, . . . , PN ← − − − − − − − − − − − − − →

select first / select last / most likely to be a true signal least likely to be a true signal

6/29

slide-7
SLIDE 7

Ordered hypothesis testing

Where does the ordering come from?

  • Data from related experiments: e.g. gene expression levels in

a different tissue, with a related drug compound, etc

  • Regression setting:

For sequential procedures (forward selection, LASSO, etc), recent work produces valid p-values for variables in the order that they are selected:

  • Post-selection inference

(Fithian, Taylor, Tibshirani, Tibshirani, Lockart, ....)

  • Knockoff method (Barber & Cand`

es): one-bit p-values

7/29

slide-8
SLIDE 8

Ordered hypothesis testing

SeqStep method (Barber & Cand` es):

  • 100

200 300 400 500 0.0 0.4 0.8 Index p−value

Want to estimate # nulls among first k p-values count how many p-values are > 0.5

8/29

slide-9
SLIDE 9

Ordered hypothesis testing

Null p-values are equally likely to be above 0.5 or below 0.5 ⇓ ≈ half the null p-values, among the first k p-values, will be > 0.5 ⇓ FDP(k) ≈ 2 · (# p-values > 0.5, among first k) k = FDPSeqStep(k) Then stop at kSeqStep = last time that FDPSeqStep(k) ≤ α

9/29

slide-10
SLIDE 10

Ordered hypothesis testing

A related method — ForwardStop (G’Sell et al 2013): To estimate FDP among the first k p-values,

  • FDPForwardStop(k) =

k

i=1 log

  • 1

1−Pi

  • k

Then stop at kForwardStop = last time that FDPForwardStop(k) ≤ α

10/29

slide-11
SLIDE 11

Accumulation tests

Accumulation test: reject the first kh p-values, where

  • kh = max
  • k :

FDPh(k) ≤ α

  • ,

for FDP(k) = # nulls among {1, . . . , k} k ≈ h(P1) + · · · + h(Pk) k

  • Estimated FDP=

FDPh(k)

h is a function [0, 1] → [0, ∞] with

  • 1

t=0 h(t) dt = 1 ⇒ E [h(Pi)] = 1 for the nulls

  • h ≈ 0 near 0 ⇒ E [h(Pi)] ≈ 0 for strong signals

11/29

slide-12
SLIDE 12

Accumulation tests

Existing & new choices for the function h:

0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

SeqStep (knockoff paper)

P h(P) 0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

ForwardStop (G'Sell et al 2013)

P h(P) 0.0 0.2 0.4 0.6 0.8 1.0 1 2 3 4

HingeExp (new)

P h(P)

12/29

slide-13
SLIDE 13

Accumulation tests

Theorem

If h is an accumulation function bounded by C, then E # nulls among {1, . . . , k} k + C/α

  • ≤ α.

(See paper for a guarantee when h is unbounded.) Advantage over BH & other multiple testing corrections: No dependence on n = # of hypotheses tested

13/29

slide-14
SLIDE 14

Gene dosage data

  • Expression levels for n = 22283 genes measured at different

dosage levels: Sample size: 5 control (zero dose), 5 low dose, 5 high dose

  • Can we identify genes with differential expression at the

lowest dosage level?

1007_s_at 121_at 1053_at 117_at 1255_g_at 2 4 6 8 10 control low dose high dose

Data from Coser et al 2003 via R Geoquery package (data set GDS2324) 14/29

slide-15
SLIDE 15

Gene dosage data

  • Standard approach w/o high dose data:
  • 1. Two-sample test for control vs. low dose
  • 2. Then correct for multiple comparisons (BH & variants)

1007_s_at 121_at 1053_at 117_at 1255_g_at 2 4 6 8 10 control low dose high dose

  • 1007_s_at

121_at 1053_at 117_at 1255_g_at 2 4 6 8 10 control low dose

  • Our approach:
  • 1. Rank genes by comparing high dose vs. control/low dose
  • 2. Run accumulation test to compare control vs. low dose

1007_s_at 121_at 1053_at 117_at 1255_g_at 2 4 6 8 10 control low dose high dose

  • 1007_s_at

121_at 1053_at 117_at 1255_g_at 2 4 6 8 10 control / low dose high dose

  • 1007_s_at

121_at 1053_at 117_at 1255_g_at 2 4 6 8 10 control low dose

15/29

slide-16
SLIDE 16

Gene dosage data

Target FDR level q # of discoveries 0.0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 5000 10000 15000 20000

  • HingeExp

SeqStep ForwardStop Variants of BH procedure (see paper for details)

Target FDR level α 16/29

slide-17
SLIDE 17

Outline

  • 1. Accumulation tests: testing a ranked list of hypotheses
  • Joint work with Ang Li
  • 2. The p-filter: FDR control across groups
  • Joint work with Aaditya Ramdas

17/29

slide-18
SLIDE 18

Structured set of hypotheses

Time 1 Time 2 Time 3 Hypotheses: Timepoint Location

18/29

slide-19
SLIDE 19

Structured set of hypotheses

  • n hypotheses with p-values P1, . . . , Pn
  • M “layers” = partitions of the hypotheses

(e.g. entries, rows, columns in our array)

  • Goal: select set

S of discoveries such that FDR is bounded simultaneously for layer 1, 2, . . . , M.

19/29

slide-20
SLIDE 20

Structured set of hypotheses

Where do the groupings come from?

  • Natural structure in the set of hypotheses
  • Regression setting:

Clusters / correlations within the features; Hierarchical structure (e.g. due to interaction terms)

20/29

slide-21
SLIDE 21

Multilayer FDR

How to define FDR for the mth layer?

  • Partition [n] = Am

1 ∪ · · · ∪ Am Gm

  • Nulls H0

m = {g : Am g ⊆ H0}

  • Selected set

Sm = {g : Am

g ∩

S = ∅}

  • FDR control: E
  • |H0

m∩

Sm| | Sm|

  • ≤ αm?

21/29

slide-22
SLIDE 22

Multilayer FDR

A naive method:

  • For the mth layer,

— Calculate Simes p-values P m

1 , . . . , P m Gm

(P m

g

tests whether group Am

g is all nulls)

— Run BH with threshold αm on this list reject groups with P m

g

≤ adaptive threshold tm

  • Problem: results might not be consistent across the M layers

22/29

slide-23
SLIDE 23

Multilayer FDR

αindiv = 0.1 αgroup = 0.2

0.03 0.01 0.18 0.04 0.08 0.05 0.11 0.06 0.01 0.89 0.14 0.12 0.58 0.11 0.11 0.88 0.24 0.09 0.66 0.45 Simes p−value Group 1 0.05 Group 2 0.05 Group 3 0.18 Group 4 0.45 23/29

slide-24
SLIDE 24

Multilayer FDR

αindiv = 0.1 αgroup = 0.2

0.03 0.01 0.18 0.04 0.08 0.05 0.11 0.06 0.01 0.89 0.14 0.12 0.58 0.11 0.11 0.88 0.24 0.09 0.66 0.45 Simes p−value Group 1 0.05 Group 2 0.05 Group 3 0.18 Group 4 0.45 23/29

slide-25
SLIDE 25

Multilayer FDR

αindiv = 0.1 αgroup = 0.2

0.03 0.01 0.18 0.04 0.08 0.05 0.11 0.06 0.01 0.89 0.14 0.12 0.58 0.11 0.11 0.88 0.24 0.09 0.66 0.45 Simes p−value Group 1 0.05 Group 2 0.05 Group 3 0.18 Group 4 0.45 23/29

slide-26
SLIDE 26

Multilayer FDR

The p-filter:

S(t1, . . . , tm) = set of discoveries at thresholds t1, . . . , tM: Pi is selected, if it belongs to a selected group in all M layers

  • Now estimate FDP’s for

S(t1, . . . , tm), in each layer:

  • FDPm =

tm · Gm | Sm(t1, . . . , tm)| ← approx. # false discoveries ← # discoveries

  • Choose tm’s adaptively: maximize tm’s s.t.

FDPm ≤ αm ∀ m.

24/29

slide-27
SLIDE 27

Theoretical results

Theorem 1

This maximum is well-defined and can be computed efficiently. Algorithm:

  • Initialize thresholds t1 = α1, . . . , tM = αM
  • Cycle through layers 1, . . . , M:

— Check if FDPm is low enough: tm · Gm | Sm(t1, . . . , tM)| ≤ αm ? — If not, reduce tm until FDPm is ≤ αm

  • ... until there are no more changes.

25/29

slide-28
SLIDE 28

Theoretical results

PRDS assumption: for each i ∈ H0, P {P ∈ increasing set | Pi = t} is an increasing function of t

Theorem 2

This procedure controls FDR for all layers: FDR for layer m = E

  • |H0

m ∩

Sm| | Sm|

  • ≤ αm · |H0

m|

Gm ∀ m. Key lemma: If f(P) is a decreasing function of P, then E ✶ {Pi ≤ f(P)} f(P)

  • ≤ 1.

26/29

slide-29
SLIDE 29

Simulation results

Layers: entries; rows; columns. Target FDR: αentries = αrows = αcolumns = 0.2

27/29

slide-30
SLIDE 30

Future work

  • Connection between ordered testing & online testing?
  • Create data-adaptive clusters?
  • An ordered testing approach for grouped hypotheses?

28/29

slide-31
SLIDE 31

Thank you!

Accumulation tests (w/ Ang Li): http://www.stat.uchicago.edu/~rina/accumulationtests.html Multi-FDR (w/ Aaditya Ramdas): http://www.stat.uchicago.edu/~rina/pfilter.html

29/29