SLIDE 1
ACMS 20340 Statistics for Life Sciences Chapter 15: Inference in - - PowerPoint PPT Presentation
ACMS 20340 Statistics for Life Sciences Chapter 15: Inference in - - PowerPoint PPT Presentation
ACMS 20340 Statistics for Life Sciences Chapter 15: Inference in Practice Inference in Practice Recall the simple conditions for inference about a population mean: Known standard deviation Sample obtained randomly Normally
SLIDE 2
SLIDE 3
The source of our data matters
Inference via confidence intervals and hypotheses tests depend on the samples being random (since, for instance, we treat a sample statistic as a random variable). If our data don’t come from a random sample or a randomized experiment, we have a greater chance of drawing the wrong conclusion.
SLIDE 4
Some pitfalls
Given a sample, we need to be cautious: It may be hard to tell if it is an SRS.
◮ Nonresponse or dropouts from an experiment ◮ Confidence intervals and hypothesis testing may not work if
- ur sample is given by a method more complicated than an
SRS.
◮ We have to deal with voluntary response surveys, uncontrolled
experiments, biased samples, etc.
SLIDE 5
The assumption of Normality
The underlying distribution of the population is less of an issue: many statistical procedures are based on the Normality of the sampling distribution. The Central Limit Theorem justifies our assumption. One worry: Inference procedures based on sampling distributions can be influenced by outliers.
SLIDE 6
To cut a long story short. . .
We should always plot our data to check to see if it is roughly Normal before making any inferences. In case there are outliers, or the population is strongly non-Normal, there are alternative methods that don’t require Normality and are not sensitive to outliers.
SLIDE 7
z procedures
Inference via confidence intervals and hypothesis testing are sometimes referred to as z procedures, as both start with the
- ne-sample z statistic and both use the standard Normal
distribution. Let’s briefly consider the behavior of these z procedures in practice.
SLIDE 8
Confidence Intervals
The ideal situation: High confidence and a small margin of error. High confidence means our method almost always gives the correct answers. A small margin of error means we’ve pinned down the parameter to a high degree of precision.
SLIDE 9
How to get a small margin of error?
- 1. A smaller critical value z∗ (which means a lower confidence
level).
- 2. A smaller standard deviation (which means there is less
variation in the population).
- 3. A larger sample size n (which allows for more precise
estimates).
SLIDE 10
Several important points
The margin of error only accounts for sampling error (variation due to repeated sampling, as captured by the sampling distribution). More serious difficulties: undercoverage, nonresponse. Margin of error doesn’t take these into account.
SLIDE 11
In sum. . .
The margin of error in a confidence interval ignores everything except the sample-to-sample variation due to choosing the sample randomly.
SLIDE 12
Significance Tests
We use a test of significance to describe the degree of evidence provided by a sample against the null hypothesis. More precisely, the p-value gives the degree of evidence. How small a p-value is convincing evidence against a null hypothesis?
SLIDE 13
How small a p-value?
The answer depends on two circumstances:
- 1. How plausible is the null hypothesis? (If H0 is widely
accepted, then strong evidence and thus a small p is needed.)
- 2. What are the consequences of rejecting the null hypothesis?
(Would it require an expensive change?)
SLIDE 14
Significance and alternative hypotheses
The p-value for a one-sided test is half the p-value for the two-sided test of the same null hypothesis based on the same data. The evidence against a null hypothesis is stronger when the alternative hypothesis is one-sided, since it’s based on the data plus information about the direction of possible deviations.
SLIDE 15
More on significance 1
Sample size affects statistical significance: Because large random samples have small chance variation, very small population effects can be highly significant if the sample is large. z = ¯ x − µ0 σ/√n = size of the observed effect size of chance variation
SLIDE 16
More on significance 2
Because small random samples have a lot of chance variation, even large population effects can fail to be significant if the sample is small. Statistical significance does not tell us whether an effect is large enough to be important. In other words, statistical significance is not the same thing as practical significance.
SLIDE 17
Planning studies
The practical questions we can ask when planning a study are:
◮ How large should our sample size be to ensure a small margin
- f error in confidence intervals?
◮ How large should our sample size be in performing tests of
significance?
SLIDE 18
High confidence + small margin of error
We can have both a high level of confidence and a small margin of error as long as our sample is large enough. To get a margin of error: m = z∗ σ √n we need a sample of size n = z∗σ m 2 .
SLIDE 19
A familiar example
In the previous chapter, we considered mean body temperature. Population standard deviation is σ = 0.6◦F. We want to estimate mean body temperature µ for healthy adults within ±0.05◦F with 95% confidence.
SLIDE 20
The solution
The desired margin of error is m = 0.05. For 95% confidence, we have z∗ = 1.96. n = z∗σ m 2 = 1.96 × 0.6 0.05 2 = 553.2
SLIDE 21
Sample size in significance tests
How large of a sample should we take? Worry: If our sample is too small, large effects in the population might fail to give statistically significant results.
SLIDE 22
Three Questions
We must answer the following to decide how large a sample to take: Significance How much protection do we want against level: getting a significant result from our sample when there really is no effect in the population? Effect size: How large an effect in the population is important in practice? Power: How confident do we want to be that our study will detect an effect of the size we think is important?
SLIDE 23
Power
Suppose that we determine an effect size that is important in practice. The probability that our test successfully detects an effect of the specified size is the power of the test. The higher the power of a test, the more sensitive it is to deviations from the null hypothesis.
SLIDE 24
An illustration (I)
Suppose we are performing a hypothesis test with the following null and alternative hypotheses: H0 : µ = 0 Ha : µ > 0 Suppose further that an effect µ > 0.8 has practical importance for us.
SLIDE 25
An illustration (II)
We want to ensure that our test will reject the null hypothesis if the effect µ > 0.8 really is true. We can’t be 100% certain that this will happen. The power of our test is the probability that we will reject the null hypothesis when this effect really does occur.
SLIDE 26
Two probabilities (I)
We can assess the performance of a test by giving two probabilities:
- 1. The significance level α.
- 2. The power for an alternative that we want to detect.
SLIDE 27
Two probabilities (II)
The significance level of a test is the probability of making the wrong decision when the null hypothesis is true. The power against a specific alternative is the probability of making the right decision when that alternative is true.
SLIDE 28
Type I and Type II Errors
If we reject H0 when in fact H0 is true, this is a Type I error. If we fail to reject H0 when in fact Ha is true, this is a Type II error.
SLIDE 29
The probability of error
The significance level α of any fixed-level test is the probability of a Type I error. The probability of a Type II error is denoted β. The power of a test against any alternative is 1 − β.
SLIDE 30