Estimating the number and effect sizes of non-null hypotheses
Jennifer Brennan, Ramya Korlakai Vinayak, Kevin Jamieson jrb@cs.washington.edu ICML 2020
sizes of non-null hypotheses Jennifer Brennan, Ramya Korlakai - - PowerPoint PPT Presentation
Estimating the number and effect sizes of non-null hypotheses Jennifer Brennan, Ramya Korlakai Vinayak, Kevin Jamieson jrb@cs.washington.edu ICML 2020 Example: Fruit Fly Genetics Hao et al. (2008) measured the effect of 13,000 fruit fly genes
Jennifer Brennan, Ramya Korlakai Vinayak, Kevin Jamieson jrb@cs.washington.edu ICML 2020
Hao et al. (2008) measured the effect of 13,000 fruit fly genes on susceptibility to influenza Measurements were distributed N(0,1) under the null, higher indicates protection from influenza
More protection from influenza
Hao et al. (2008) measured the effect of 13,000 fruit fly genes on susceptibility to influenza Measurements were distributed N(0,1) under the null, higher indicates protection from influenza
Multiple hypothesis testing identifies few discoveries
Significant Genes
Hao et al. (2008) measured the effect of 13,000 fruit fly genes on susceptibility to influenza Measurements were distributed N(0,1) under the null, higher indicates protection from influenza
Observed distribution does not match theoretical null
π 0, 1
Hao et al. (2008) measured the effect of 13,000 fruit fly genes on susceptibility to influenza Measurements were distributed N(0,1) under the null, higher indicates protection from influenza
Observed distribution does not match theoretical null
π 0, 1
Too many small, positive measurements for chance alone
Hao et al. (2008) measured the effect of 13,000 fruit fly genes on susceptibility to influenza Measurements were distributed N(0,1) under the null, higher indicates protection from influenza
Observed distribution does not match theoretical null
π 0, 1
Too small to claim individual significance
Idea: These genes can be counted, even though they canβt be identified
Idea: These genes can be counted, even though they canβt be identified
>7% of genes have effect size >1/4 (at least 8% increase in influenza resistance)
Our Estimator
Idea: These genes can be counted, even though they canβt be identified
>2% of genes have effect size >1 (at least 28% increase in influenza resistance) >7% of genes have effect size >1/4 (at least 8% increase in influenza resistance)
Our Estimator
Idea: These genes can be counted, even though they canβt be identified Enables power analysis for future experimental designs
>2% of genes have effect size >1 (at least 28% increase in influenza resistance) >7% of genes have effect size >1/4 (at least 8% increase in influenza resistance)
Our Estimator
Idea: These genes can be counted, even though they canβt be identified
Next Experiment: Take precise measurements (e.g., use many replications) to identify these genes
Enables power analysis for future experimental designs
>2% of genes have effect size >1 (at least 28% increase in influenza resistance) >7% of genes have effect size >1/4 (at least 8% increase in influenza resistance)
Our Estimator
Idea: These genes can be counted, even though they canβt be identified
Next Experiment: Take precise measurements (e.g., use many replications) to identify these genes Next Experiment: Take less precise measurements, identify fewer genes
Enables power analysis for future experimental designs
>2% of genes have effect size >1 (at least 28% increase in influenza resistance) >7% of genes have effect size >1/4 (at least 8% increase in influenza resistance)
Our Estimator
We view multiple hypothesis testing from the perspective of learning mixture distributions
We view multiple hypothesis testing from the perspective of learning mixture distributions For π = 1, 2, β¦ , π Draw ππ βΌ πβ
We view multiple hypothesis testing from the perspective of learning mixture distributions For π = 1, 2, β¦ , π Draw ππ βΌ πβ ππ is the (unknown) effect size
We view multiple hypothesis testing from the perspective of learning mixture distributions For π = 1, 2, β¦ , π Draw ππ βΌ πβ ππ is the (unknown) effect size Observe ππ βΌ π(ππ)
We view multiple hypothesis testing from the perspective of learning mixture distributions For π = 1, 2, β¦ , π Draw ππ βΌ πβ ππ is the (unknown) effect size Observe ππ βΌ π(ππ) E.g. π ππ = π(ππ, 1)
We view multiple hypothesis testing from the perspective of learning mixture distributions For π = 1, 2, β¦ , π Draw ππ βΌ πβ ππ is the (unknown) effect size Observe ππ βΌ π(ππ) E.g. π ππ = π(ππ, 1) Identification: Which ππ > 0? Counting: What is the probability π
πβΌπβ(π > 0)?
We view multiple hypothesis testing from the perspective of learning mixture distributions For π = 1, 2, β¦ , π Draw ππ βΌ πβ ππ is the (unknown) effect size Observe ππ βΌ π(ππ) E.g. π ππ = π(ππ, 1) Identification: Which ππ > 0? Counting: What is the probability π
πβΌπβ(π > πΏ), for all πΏ?
We view multiple hypothesis testing from the perspective of learning mixture distributions For π = 1, 2, β¦ , π Draw ππ βΌ πβ ππ is the (unknown) effect size Observe ππ βΌ π(ππ) E.g. π ππ = π(ππ, 1) Identification: Which ππ > 0? (Returns a set in [n]) Counting: What is the probability π
πβΌπβ(π > πΏ), for all πΏ?
(Returns a fraction)
We view multiple hypothesis testing from the perspective of learning mixture distributions For π = 1, 2, β¦ , π Draw ππ βΌ πβ ππ is the (unknown) effect size Observe ππ βΌ π(ππ) E.g. π ππ = π(ππ, 1) Goal Estimate ππβ πΏ = π
πβΌπβ(π > πΏ), for all πΏ
Constraint Never overestimate the true fraction
Estimating the number of non-nulls (π β 0)
Early techniques [Schweder and SpjΓΈtvoll, 1982; Genovese et al., 2004; Meinshausen et al., 2006] relied on uniformity of p-values under the null Techniques do not extend to arbitrary thresholds (βHow many genes improved influenza resistance by at least 20%?β)
Plug-in estimators
Estimate the entire density π, then compute π
π(π > πΏ)
Does not respect our constraint, that we cannot overestimate
Connections to False Discovery Rate (FDR) control
Tighter FDR control can be obtained by knowing number of non-nulls Previous methods either do not satisfy our constraint [Storey, 2002; Li and Barber,
2019], or perform poorly in our regime of interest (many hypotheses, small
effect sizes) [Stephens, 2016; Katsevich and Ramdas, 2018]
Estimating the number of non-nulls (π β 0)
Early techniques [Schweder and SpjΓΈtvoll, 1982; Genovese et al., 2004; Meinshausen et al., 2006] relied on uniformity of p-values under the null Techniques do not extend to arbitrary thresholds (βHow many genes improved influenza resistance by at least 20%?β)
Plug-in estimators
Estimate the entire density π, then compute π
π(π > πΏ)
Does not respect our constraint, that we cannot overestimate
Connections to False Discovery Rate (FDR) control
Tighter FDR control can be obtained by knowing number of non-nulls Previous methods either do not satisfy our constraint [Storey, 2002; Li and Barber,
2019], or perform poorly in our regime of interest (many hypotheses, small
effect sizes) [Stephens, 2016; Katsevich and Ramdas, 2018]
Estimating the number of non-nulls (π β 0)
Early techniques [Schweder and SpjΓΈtvoll, 1982; Genovese et al., 2004; Meinshausen et al., 2006] relied on uniformity of p-values under the null Techniques do not extend to arbitrary thresholds (βHow many genes improved influenza resistance by at least 20%?β)
Plug-in estimators
Estimate the entire density π, then compute π
π(π > πΏ)
Does not respect our constraint, that we cannot overestimate
Connections to False Discovery Rate (FDR) control
Tighter FDR control can be obtained by knowing number of non-nulls Previous methods either do not satisfy our constraint [Storey, 2002; Li and Barber,
2019], or perform poorly in our regime of interest (many hypotheses, small
effect sizes) [Stephens, 2016; Katsevich and Ramdas, 2018]
Goal Estimate Constraint Never overestimate Step 1 Consider the empirical CDF (Cumulative Distribution Function)
Step 1 Consider the empirical CDF (Cumulative Distribution Function) Step 2 Generate confidence intervals on the true CDF Goal Estimate Constraint Never overestimate
DKW Inequality
Step 1 Consider the empirical CDF (Cumulative Distribution Function) Step 2 Generate confidence intervals on the true CDF Goal Estimate Constraint Never overestimate
With high probability, the true CDF lives within this interval DKW Inequality
Goal Estimate Constraint Never overestimate Step 1 Consider the empirical CDF (Cumulative Distribution Function) Step 2 Generate confidence intervals on the true CDF
With high probability, the true CDF lives within this interval DKW Inequality
Goal Estimate Constraint Never overestimate Step 1 Consider the empirical CDF (Cumulative Distribution Function) Step 2 Generate confidence intervals on the true CDF
With high probability, the true CDF lives within this interval This could be the true CDF N(0,1) could not be the true CDF
Step 1 Consider the empirical CDF (Cumulative Distribution Function) Step 2 Generate confidence intervals on the true CDF Step 3 Return the smallest amount of mass that could feasibly have generated the empirical CDF Goal Estimate Constraint Never overestimate
With high probability, the true CDF lives within this interval This is the CDF in this interval with the least mass above Ξ³
Step 1 Consider the empirical CDF (Cumulative Distribution Function) Step 2 Generate confidence intervals on the true CDF Step 3 Return the smallest amount of mass that could feasibly have generated the empirical CDF Goal Estimate Constraint Never overestimate
A convex program! (efficiently computable) Constraint is satisfied with high probability Our sample complexity results match known lower bounds
Step 1 Consider the empirical CDF (Cumulative Distribution Function) Step 2 Generate confidence intervals on the true CDF Step 3 Return the smallest amount of mass that could feasibly have generated the empirical CDF Goal Estimate Constraint Never overestimate
A convex program! (efficiently computable) Constraint is satisfied with high probability Our sample complexity results match known lower bounds
Step 1 Consider the empirical CDF (Cumulative Distribution Function) Step 2 Generate confidence intervals on the true CDF Step 3 Return the smallest amount of mass that could feasibly have generated the empirical CDF Goal Estimate Constraint Never overestimate
Our sample complexity results match known lower bounds A convex program! (efficiently computable) Constraint is satisfied with high probability
Our estimator provides the following guarantees:
With probability 1 β π½, does not overestimate ππβ πΏ for any πΏ
For π = 1, 2, β¦ , π Draw ππ βΌ πβ Observe ππ βΌ π ππ Estimate ππβ πΏ = π
πβΌπβ(π > πΏ)
Our estimator provides the following guarantees:
With probability 1 β π½, does not overestimate ππβ πΏ for any πΏ With probability 1 β π, estimate is at most π» from the truth whenever
For π = 1, 2, β¦ , π Draw ππ βΌ πβ Observe ππ βΌ π ππ Estimate ππβ πΏ = π
πβΌπβ(π > πΏ)
Our estimator provides the following guarantees:
With probability 1 β π½, does not overestimate ππβ πΏ for any πΏ With probability 1 β π, estimate is at most π» from the truth whenever
πΊ
πβ
CDFs πΊ
π corresponding to
all mixing distributions π with less than ππβ πΏ β π probability mass above πΏ For π = 1, 2, β¦ , π Draw ππ βΌ πβ Observe ππ βΌ π ππ Estimate ππβ πΏ = π
πβΌπβ(π > πΏ)
Our estimator provides the following guarantees:
With probability 1 β π½, does not overestimate ππβ πΏ for any πΏ With probability 1 β π, estimate is at most π» from the truth whenever
πΊ
πβ
Minimum ββ distance CDFs πΊ
π corresponding to
all mixing distributions π with less than ππβ πΏ β π probability mass above πΏ For π = 1, 2, β¦ , π Draw ππ βΌ πβ Observe ππ βΌ π ππ Estimate ππβ πΏ = π
πβΌπβ(π > πΏ)
Our estimator provides the following guarantees:
With probability 1 β π½, does not overestimate ππβ πΏ for any πΏ With probability 1 β π, estimate is at most π» from the truth whenever
πΊ
πβ
Minimum ββ distance CDFs πΊ
π corresponding to
all mixing distributions π with less than ππβ πΏ β π probability mass above πΏ
Goal: lower bound this distance
For π = 1, 2, β¦ , π Draw ππ βΌ πβ Observe ππ βΌ π ππ Estimate ππβ πΏ = π
πβΌπβ(π > πΏ)
Let ππ~π(ππ, 1) be drawn from a mixture of Gaussians, with πβ alternate hypotheses of effect size πΏβ < 1 With probability at least 1 β π, our estimator detects over half of the alternate hypotheses (i.e., α ππ 0 >
1 2 πβ), whenever
Matches a novel lower bound
πΏβ πβ How much mass is (strictly) above 0?
What fraction of genes are non-null?
What fraction of genes are non-null? What fraction of genes have effect at least Β½ the true alternate?
What fraction of genes are non-null? What fraction of genes have effect at least Β½ the true alternate?
What fraction of genes are non-null? What fraction of genes have effect at least Β½ the true alternate?
Standardized Testing
Each school has some ππ indicating its studentsβ true performance We observe ππ, a noisy measurement of ππ (e.g., studentsβ average exam score) Our estimator: βat least Y% of schools are below proficient in mathβ Interesting on its own, or to suggest further testing to identify these schools
Standardized Testing
Each school has some ππ indicating its studentsβ true performance We observe ππ, a noisy measurement of ππ (e.g., studentsβ average exam score) Our estimator: βat least Y% of schools are below proficient in mathβ Interesting on its own, or to suggest further testing to identify these schools
Public Health*
Each person has some ππ indicating their susceptibility to the flu (variable due to age, health, etc.) We observe ππ, the number of flu seasons they were sick, in the past five years Our estimator: βat most Y% of people have a 25% chance or greater of getting sick in a given yearβ (Impossible to identify these people with confidence)
*Example due to Tian, Kong and Valiant (2017)