Unit 7: Inference for Quantitative Data: Means
Why Inference for Means Uses the t Distribution
Parameters, statistics, and the problem we’re solving
Inference is about using data from a sample to make a justified claim about a population. For quantitative data (numbers you can average), the population feature you usually care about is the population mean, written as . Because you almost never know , you collect a sample and compute the sample mean as an estimate.
The key challenge is uncertainty: changes from sample to sample. This unit is about measuring that uncertainty when your parameter is a mean, and using it to build:
- Confidence intervals (estimate with a range of plausible values)
- Significance tests (evaluate whether data provide convincing evidence against a claim about )
Why we can’t usually use the normal (z) procedures
If you somehow knew the population standard deviation , then the standardized statistic
would follow a standard normal model (approximately, under appropriate conditions). But in real life, is almost never known. Instead, you estimate it with the sample standard deviation .
That substitution changes the sampling behavior: when you estimate variability from the same sample you’re using to estimate the mean, you add extra uncertainty. That extra uncertainty is exactly why we use Student’s t distribution instead of the normal distribution.
The t distribution: what it is, where it came from, and how it behaves
The Student’s t-distribution was introduced in 1908 by W. S. Gosset, a British mathematician employed by the Guinness Breweries.
A t distribution is a family of bell-shaped, symmetric distributions centered at 0. Like the normal, it’s used for standardized statistics, but it has heavier tails and is a bit lower near the mean. Heavier tails matter because they assign more probability to extreme values. This reflects the fact that itself fluctuates from sample to sample, so your standardized statistic is “less stable” than .
When doing inference for means with unknown , the standardized statistic becomes:
If the population is normally distributed, then this statistic follows a t distribution.
The exact t distribution you use depends on the degrees of freedom (df). For a one-sample mean procedure,
The smaller the df value, the more spread out the t distribution is. As gets larger (df increases), the t distribution gets closer and closer to the standard normal. In practice:
- Small : t has noticeably heavier tails (larger critical values, wider intervals)
- Large : t is very close to z
Because is almost always unknown in the real world, t procedures are the default choice for inference about means.
When t procedures are valid: the conditions you must check
AP Statistics emphasizes that inference procedures are only trustworthy when certain conditions are met. For t procedures about means, think in three categories.
1) Random: Data come from a random sample or a randomized experiment.
- Random sampling supports generalizing to the population.
- Random assignment (in an experiment) supports cause-and-effect conclusions.
2) Independence: Observations are independent.
- A common check for sampling without replacement is the 10% condition: the sample size should be no more than 10% of the population size.
3) Normal (or large sample): The sampling distribution of (or of differences) is approximately normal.
- If the population distribution is roughly normal, you’re fine even with small .
- If the population is not normal, you typically want a “large enough” sample so the Central Limit Theorem makes approximately normal.
- A common rule of thumb is for CLT-based reasoning, but strong skewness or outliers can still cause trouble.
A practical way to handle the Normal condition is to look at graphs (dotplot, histogram, boxplot) and ask: is the distribution roughly symmetric and free of extreme outliers? If not, do you have a large sample size to rely on the CLT?
Exam Focus
- Typical question patterns:
- Explain why t procedures are used instead of z procedures when is unknown.
- Identify the parameter and state the appropriate df for a given situation.
- Check conditions (Random, 10%, Normal/large sample) from a description and/or graph.
- Common mistakes:
- Using z critical values (or saying “normal”) when the problem clearly indicates is unknown.
- Forgetting the 10% condition when sampling without replacement.
- Treating “n is kind of big” as automatic permission even when there are extreme outliers.
One-Sample t Confidence Intervals for a Population Mean
What a confidence interval means (and what it doesn’t)
A confidence interval for is a range of values that are plausible for the true population mean, based on your sample. The logic is:
- The sample mean is your best point estimate of .
- The sample-to-sample variation in is summarized by the standard error.
- You create an interval by taking and moving out by a margin of error.
The key interpretation (the one AP wants) is about the long-run performance of the method:
If we repeatedly took random samples of size from the same population and built a t interval each time, then about % of those intervals would capture the true mean .
A very common misconception is to say “there is a % probability that is in this interval.” After you compute an interval, is fixed; the interval either contains it or not. The confidence level describes the method, not the one interval.
The sampling distribution ideas behind the interval
Your sample mean is just one of a whole universe of sample means. If is sufficiently large (or if the population is normally distributed), then:
- The set of all sample means is approximately normally distributed.
- The mean of the set of sample means equals , the population mean.
- The standard deviation of the set of sample means is
Because we typically do not know , we estimate it using . In that case, the estimated standard deviation of is the standard error:
The structure of a one-sample t interval
The one-sample t confidence interval for is:
Here’s what each piece does:
- : center of the interval (your estimate)
- : standard error of
- : critical t value based on the confidence level and
The margin of error is:
Why sample size and confidence level matter
The standard error shrinks like , so larger samples give more precise estimates.
If you increase :
- tends to decrease
- df increases, making smaller
- the interval gets narrower
If you raise the confidence level (say 90% to 95%):
- increases
- margin of error increases
- the interval gets wider
Example: building and interpreting a one-sample t interval
Scenario. A random sample of customers is surveyed about the amount of time (in minutes) they spend in a store. The sample mean is minutes and the sample standard deviation is minutes. Construct and interpret a 95% confidence interval for the true mean time .
Step 1: Identify the parameter.
- = the true mean time spent in the store by all customers in the population of interest.
Step 2: Check conditions.
- Random: stated as a random sample.
- Independence: assume the population is at least 10 times 25.
- Normal/large sample: with , you’d want the distribution to be roughly symmetric with no extreme outliers.
Step 3: Compute the interval.
For 95% confidence and , .
So:
Interval endpoints:
- Lower:
- Upper:
Rounded reasonably: minutes.
Step 4: Interpret in context.
We are 95% confident that the true mean time that all customers spend in the store is between about 69.1 and 75.7 minutes.
Example 7.1: gas mileage confidence intervals and “what confidence is this margin?”
When a random sample of 10 cars of a new model was tested for gas mileage, the results showed a mean of 27.2 miles per gallon with a standard deviation of 1.8 miles per gallon.
1) What is a 95% confidence interval estimate for the mean gas mileage achieved by this model? Assume the population of mpg results for all new-model cars is approximately normally distributed.
Parameter: Let represent the mean gas mileage (mpg) in the population of cars of this new model.
Procedure: One-sample t-interval for a population mean.
Checks: Random sample is stated, is assumed less than 10% of all such cars, and the population is approximately normal.
Mechanics (technology): A calculator t-interval gives .
Conclusion: We are 95% confident that the true mean gas mileage is between 25.91 and 28.49 mpg.
2) Based on this confidence interval, is the true mean mileage significantly different from 25 mpg?
Yes. Because 25 is not in the interval of plausible values (about 25.9 to 28.5), there is convincing evidence that the true mean mileage differs from 25 mpg.
3) Determine a 99% confidence interval.
Mechanics (technology): .
Conclusion: We are 99% confident the true mean mpg is between 25.35 and 29.05 mpg. Notice the higher confidence produces a wider interval (less specific).
4) What would the 95% confidence interval be if the same sample mean of 27.2 and standard deviation of 1.8 had come from a sample of 20 cars?
Mechanics (technology): .
Conclusion: We are 95% confident the true mean mpg is between 26.36 and 28.04 mpg. Notice the larger sample size produces a narrower interval.
5) With the original data, with what confidence can we assert that the true mean gas mileage is ?
The interval is . Convert the margin of error to a t critical value using .
With , the central area between and is about 0.899 (using a calculator t CDF), so the confidence level is about 89.9%.
What can go wrong with one-sample t intervals
Confusing and is a major issue: t intervals use because is unknown. Also, ignoring outliers with small samples can be disastrous because one extreme outlier can dramatically affect both and . Finally, df errors are common: for a one-sample interval, .
Exam Focus
- Typical question patterns:
- Construct a one-sample t interval from summary statistics or calculator output.
- Interpret a confidence interval correctly in context (including what represents).
- Explain how changing or confidence level affects the margin of error.
- Use an interval to argue whether a particular value (like 25 mpg) is plausible.
- Determine what confidence level corresponds to a given “estimate ± margin of error.”
- Common mistakes:
- Saying “95% of data fall in the interval” instead of “95% confident about .”
- Using instead of .
- Reporting an interval for individuals rather than for the population mean.
One-Sample t Tests for a Population Mean
What a significance test is really asking
A significance test uses sample data to evaluate a claim about a population parameter. For a one-sample mean test, the parameter is and the hypotheses look like:
and one of these alternatives:
A test asks: if were true, would results like ours be unusually far from ?
Conditions (with the common AP rule-of-thumb)
To conduct a significance test for a mean, check:
- Random: a simple random sample (or randomized experiment).
- Independence: the sample is less than 10% of the population when sampling without replacement.
- Normal/large sample: either the population is approximately normal, or the sample is large enough for the CLT to apply (often summarized as ). If raw data are shown, a plot should be unimodal and reasonably symmetric with no outliers and no strong skewness.
The one-sample t test statistic and p-value
The test statistic is
The p-value is the probability, assuming is true, of getting a test statistic at least as extreme as the one observed (in the direction(s) of ). A small p-value is evidence against .
Connecting p-values, significance level, and conclusions
Often you compare the p-value to a chosen significance level (commonly 0.05):
- If : reject
- If : fail to reject
“Fail to reject” is not the same as “accept.” It means you do not have strong evidence against ; it does not prove is true.
Example: one-sample t test from summary statistics
Using the store-time sample: , , .
Claim to test. A manager claims the true mean time is 70 minutes. Test at :
Compute the test statistic:
Two-sided p-value is about 0.14 to 0.15.
Decision and conclusion: Because , fail to reject . At the 5% level, the sample does not provide convincing evidence that the true mean time differs from 70 minutes.
Example 7.2: air-conditioning electricity use (and error type)
A manufacturer claims that a new brand of air-conditioning units uses only 6.5 kilowatts of electricity per day. A consumer agency believes the true figure is higher and runs a test on a random sample of size 50. If the sample mean is 7.0 kilowatts with a standard deviation of 1.4, should the manufacturer’s claim be rejected at a significance level of 5%? Of 1%? Then, given the conclusion, what type of error might have been committed and what is a possible consequence?
Parameter: Let be the mean electricity usage (kilowatts per day) for the population of these units.
Hypotheses:
Checks: Random sample is stated. Independence via 10% condition is reasonable. Since , the CLT supports approximate normality of .
Mechanics (technology): and .
Conclusions:
- At , since , reject . There is convincing evidence that the true mean usage is higher than 6.5 kW/day.
- At , since , also reject .
Error type and consequence: Because the decision was to reject , the possible mistake is a Type I error (rejecting a true null). A possible consequence is discouraging customers from purchasing a unit that really does meet the advertised electricity usage.
The confidence interval connection (powerful on exams)
There is a deep link between two-sided tests at level and confidence intervals at level :
- If a 95% confidence interval for does not contain , then a two-sided test at would reject .
- If the interval does contain , the test would fail to reject.
What can go wrong with one-sample t tests
Wrong-tail errors happen when the direction of doesn’t match the p-value you report (one-sided vs two-sided). Another common issue is misinterpreting the p-value as “the probability the null is true.” Finally, statistical significance is not the same as practical importance: with large samples, tiny differences can become statistically significant.
Exam Focus
- Typical question patterns:
- Write hypotheses for a claim about a mean and perform a one-sample t test.
- Interpret a p-value in context.
- Decide at different significance levels (for example, 5% and 1%).
- Use a confidence interval to make a test decision (or explain the relationship).
- Identify a plausible Type I or Type II error after a decision.
- Common mistakes:
- Writing (the null must include equality).
- Concluding “reject , so is true” without evidence language.
- Forgetting to define in context.
Two-Sample t Procedures for the Difference Between Two Means
When you need a two-sample method
Use two-sample t procedures when you are comparing the means of two independent groups, such as two separate random samples (men vs women) or two treatment groups in a randomized experiment. Independence is crucial: the individuals in one group are not naturally linked to individuals in the other group.
Parameters, statistics, and the sampling distribution background
Let be the true mean of population 1 and the true mean of population 2. The parameter of interest is often:
Your statistic is:
For large enough samples (or normal populations), the set of all differences of sample means is approximately normally distributed, with mean and standard deviation:
When population standard deviations are unknown, you estimate with sample standard deviations:
Two-sample t confidence interval
A two-sample t interval for is:
Degrees of freedom are typically computed by technology using the Welch-Satterthwaite approximation:
On the AP Exam, you generally are not required to compute df by hand for two-sample t procedures, but you should recognize that df depends on , , , and , and that smaller samples produce smaller df and larger critical values.
Two-sample t test
For ,
More generally, for ,
Conditions for two-sample t inference
Check conditions for each group:
1) Random
- Two random samples, or random assignment to two treatments.
2) Independence
- Within each group, observations are independent (10% condition if sampling without replacement).
- Between groups, the two samples/groups are independent.
3) Normal/large sample
- Each group’s distribution should be approximately normal, or each sample size should be large enough for the CLT (often summarized as and ).
- Outliers in either group can cause trouble, especially with small samples.
Example: two-sample t confidence interval (teaching methods)
A study compares test scores for two teaching methods.
- Method 1: , ,
- Method 2: , ,
Construct a 95% confidence interval for (Method 1 minus Method 2).
Difference in sample means:
Standard error:
Using technology, df is around the 70s, and for 95% confidence .
Interval:
So .
Interpretation: We are 95% confident that the true mean score for Method 1 is between about 0.7 points lower and 8.3 points higher than the true mean score for Method 2. Because 0 is inside the interval, a difference of 0 is plausible.
Example: two-sample t test (teaching methods)
Test:
With df around the 70s, the two-sided p-value is around 0.10. At , fail to reject . The data do not provide convincing evidence that the two teaching methods have different mean test scores.
Example 7.3: accidents per month (two-sample t interval)
A 30-month study is conducted to determine the difference in the numbers of accidents per month occurring in two departments in an assembly plant. Suppose the first department averages 12.3 accidents per month with a standard deviation of 3.5, while the second averages 7.6 accidents with a standard deviation of 3.4. Determine a 95% confidence interval for the mean difference (first department minus second department). Assume the populations are independent and approximately normally distributed.
Parameters:
- = mean accidents per month for the first department (across all months of interest)
- = mean accidents per month for the second department (across all months of interest)
Procedure: Two-sample t-interval for .
Checks: The setup does not explicitly state random sampling, so you would need to assume the 30 months are representative. Independence between departments is stated, and approximate normality is stated.
Mechanics (technology): .
Conclusion: We are 95% confident that the first department has a true mean of between 2.92 and 6.48 more accidents per month than the second department.
Significance idea: Yes, because the entire interval is positive, there is convincing evidence that the mean difference is not 0.
Example 7.4: difference of two means test (downtime) plus Type II error
A sales representative believes his company’s computers have more average non-operational time per week than a competitor’s model. He takes two independent simple random samples.
- Company: , minutes,
- Competitor: , minutes,
Parameters: = mean downtime for the company’s computer; = mean downtime for the competitor’s computer.
Hypotheses (matching the claim “company is higher”):
Checks: Independent SRSs are given. Since and , CLT conditions are reasonable.
Mechanics (technology): and .
Conclusion: Because , fail to reject . There is not convincing evidence that the company’s computers have greater mean downtime than the competitor’s.
Error type and consequence: Since the decision was to fail to reject , the possible mistake is a Type II error (failing to reject a false null). A possible consequence is that the company’s computers truly have higher downtime, but the test doesn’t detect it, so no fixes are made and future sales suffer.
What can go wrong with two-sample t procedures
A common conceptual error is using two-sample methods for paired data. Another is assuming equal variances without justification; standard AP technology routines typically use unpooled (Welch) procedures. Finally, be careful about direction: means “group 1 minus group 2,” and your sign and interpretation must match that order.
Exam Focus
- Typical question patterns:
- Choose between paired t and two-sample t and justify using the study design.
- Compute and interpret a two-sample t interval for .
- Perform a two-sample t test and write a full conclusion in context.
- Use whether 0 is in the interval to discuss “significant difference.”
- Identify a plausible Type II error when you fail to reject.
- Common mistakes:
- Checking Normality only once instead of for both groups.
- Stating conclusions about sample means instead of population means.
- Mixing up which mean is subtracted from which and interpreting the sign backwards.
Paired t Procedures (Matched Pairs) for Mean Differences
When data are paired and why that changes the analysis
Not all “two measurements” problems are two-sample problems. Sometimes each individual provides two related measurements: before/after, left/right hand, twin pairs, or a carefully matched pair.
In these situations, the two observations within a pair are not independent. If you incorrectly treat them as independent samples, you will usually underestimate variability and produce misleading results.
The matched-pairs strategy is to reduce the problem to one quantitative variable: the difference within each pair.
The parameter for paired t inference
Define a difference for each pair, usually as:
Then your parameter becomes:
- : the true mean of the differences in the population of pairs
You compute:
- : the sample mean of the differences
- : the sample standard deviation of the differences
- : the number of pairs
Once you have the differences, you perform a one-sample t interval or t test on the differences.
Paired t confidence interval
A paired t confidence interval for is:
with
Paired t test
For testing a claim about the mean difference:
Most paired tests use , because “no average change” corresponds to a mean difference of 0.
Example: paired t test (before/after)
A researcher measures reaction time (milliseconds) of people before and after drinking coffee. Let . The sample has ms and ms. Test whether coffee reduces reaction time on average at .
Parameter and hypotheses:
- = true mean difference (after minus before)
- “Reduces reaction time” means after is smaller, so differences tend to be negative.
Conditions:
- Random: assume the people are randomly selected.
- Independence: pairs are independent of other pairs (10% condition if sampling without replacement).
- Normal/large sample: is small, so check that the differences are roughly normal with no strong outliers.
Compute:
p-value is about 0.02 (one-sided). Since , reject . There is convincing evidence that mean reaction time after coffee is lower than before coffee.
Example 7.5: SAT prep class (paired-data interval)
An SAT preparation class of 30 randomly selected students produces summary information for total score improvement (second score minus first score). Find a 90% confidence interval for the mean improvement.
Key idea: It would be wrong to calculate a two-sample interval using the two sets of scores separately, because the independence condition between the “two samples” is violated. The correct approach is to compute differences (improvements) and do a one-sample t-interval on those differences.
Parameter: Let represent the mean improvement (second minus first) for the population of students who take this class.
Procedure: One-sample t-interval for the mean of a population of differences.
Checks: Random sample is stated; is less than 10% of all such students; and is large enough for CLT reasoning.
Mechanics (using the differences): With and for the improvement variable, technology gives the 90% interval .
Conclusion: We are 90% confident that the true mean improvement in SAT scores is between 33.59 and 50.91 points.
What can go wrong with matched pairs
Forgetting to analyze differences is the biggest issue: the t procedure is on the values, not on the original two columns separately. Also, the order of subtraction matters; your hypotheses and interpretation must match your definition of . Finally, treating paired data as two independent samples is one of the most common conceptual errors.
Exam Focus
- Typical question patterns:
- Decide whether a situation is matched pairs or two-sample, and justify.
- Define the difference variable and carry out a paired t test or interval.
- Interpret the conclusion in terms of the original context (before vs after), not just “differences.”
- Common mistakes:
- Using two-sample t procedures when the same individuals are measured twice.
- Defining one way but writing hypotheses as if it were defined the opposite way.
- Checking Normality on the original measurements instead of on the differences.
Reading Technology Output and Communicating Results Clearly
Why communication matters as much as calculations
In practice, people rarely compute t procedures entirely by hand; software and calculators do the arithmetic. AP Statistics still expects you to understand what the output means and to communicate results in correct statistical language.
A strong inference response typically includes:
1) Parameter defined in context
2) Conditions checked and stated
3) Method named (one-sample t interval, paired t test, two-sample t test, etc.)
4) Key numbers (test statistic and p-value, or interval endpoints)
5) Conclusion in context, matching the hypotheses
The biggest scoring losses often come from vague or incorrect interpretations, not from minor arithmetic errors.
Common pieces of calculator/software output
You’ll commonly see:
- : the test statistic
- : degrees of freedom
- : p-value
- or and : sample mean(s)
- or and : sample standard deviation(s)
- For intervals: endpoints (lower, upper) and sometimes
Interpreting output correctly means connecting each number to the statistical story:
- tells you how far your sample result is from the null, in standard error units.
- tells you how surprising that distance would be if the null were true.
- Interval endpoints tell you what values of the parameter are plausible.
Writing conclusions that match the question
For tests, use “convincing evidence” language, reference the parameter and context, and avoid saying “proved” or “disproved.”
Example conclusion template (test):
Because the p-value of ___ is (less/greater) than \alpha=___, we (reject/fail to reject) . There is (convincing/not convincing) evidence that ___ (state in context, referring to the population and the mean).
Example conclusion template (interval):
We are ___% confident that ___ (parameter in context) lies between ___ and ___.
Rounding and reporting
AP does not demand identical rounding across all responses, but your work should be consistent and reasonable:
- Carry extra decimals during intermediate steps.
- Round final answers appropriately for the measurement context.
- If using calculator output, reporting endpoints as given is usually acceptable.
Confidence intervals versus hypothesis tests (and parameters versus statistics)
A claim about a population parameter indicates a hypothesis test, while an estimate of a population parameter asks for a confidence interval.
Also, be clear on parameters versus statistics:
- A parameter describes a population.
- A statistic describes a sample.
Example 7.8: test versus interval consistency (illustrated with proportions)
Random samples of 3-point shots by basketball players Stephen Curry and Michael Jordan show a 43% rate for Curry and a 33% rate for Jordan.
(a) Are these numbers parameters or statistics?
They are statistics because they describe samples, not all 3-point shots ever taken.
(b) State appropriate hypotheses for testing whether the difference is statistically significant.
(c) Suppose the sample sizes were both 100. Find the z-statistic and p-value, and conclude at the 5% level.
Technology gives and . Since , fail to reject . There is not convincing evidence of a difference in the true 3-point percentage rates.
(d) Calculate and interpret a 95% confidence interval for the difference in population proportions.
Technology gives . We are 95% confident that the true difference in 3-point percentage rates (Curry minus Jordan) is between −3.4% and 23.4%.
(e) Are the test decision and confidence interval consistent?
Yes. We failed to reject , and the interval includes 0.
(f) Repeat (c), (d), and (e) if the sample size had been 200 for each.
Technology gives and . Since , reject ; there is convincing evidence of a difference. The 95% interval is , which does not include 0, so it is also convincing evidence of a difference. Again, the interval and test are consistent.
What can go wrong when interpreting output
A common error is treating a confidence interval as a statement about individual values; it’s about the mean (or other parameter). Another is mismatching the alternative with the p-value (using a two-sided p-value when you set up a one-sided alternative, or vice versa). Finally, don’t ignore design limitations: random assignment allows causal conclusions, but observational comparisons do not.
Exam Focus
- Typical question patterns:
- Interpret calculator output (identify , , p-value, interval) and write a conclusion.
- Decide whether the result is statistically significant at a given and explain.
- Explain what can be concluded (and what cannot) based on sampling versus random assignment.
- Decide whether a prompt calls for a confidence interval or a hypothesis test.
- Use the CI-test consistency idea (value in interval versus reject/fail to reject).
- Common mistakes:
- Saying “there is a 95% chance is in the interval.”
- Forgetting to link the conclusion to the correct population.
- Declaring causation in an observational study.
Simulations and P-Values
Why simulations can produce (approximate) p-values
A simulation can approximate what values of a test statistic are likely to occur by random chance alone, assuming the null hypothesis is true. Then, by seeing where your observed statistic falls relative to the simulated distribution, you can estimate a p-value.
This idea is broader than t tests: it’s useful whenever the usual theoretical model is hard to justify, or when the “test statistic” isn’t one of the classic t or z statistics.
Example 7.6: simulation-based p-value using MAD
One measure of variability is the median absolute deviation (MAD). It is defined as the median of the absolute deviations from the median.
A quality control check collects random samples of size 6. If the MAD is significantly greater than expected when the machinery is operating properly, a recalibration is necessary. In a simulation of 100 samples of size 6 from proper operation, the resulting MAD values form a reference distribution.
Suppose a random sample of 6 measurements is:
{8.04, 8.06, 8.10, 8.14, 8.18, 8.19}
Compute the MAD:
- Median of the six values is the average of the 3rd and 4th values:
- Absolute deviations from 8.12 are:
{0.08, 0.06, 0.02, 0.02, 0.06, 0.07}
- The MAD is the median of these six deviations, the average of the 3rd and 4th values when sorted:
In the simulation, 3 out of 100 simulated MAD values were 0.06 or greater, so the estimated p-value is 0.03.
Decision: With , there is sufficient evidence to necessitate a recalibration of the machinery.
Exam Focus
- Typical question patterns:
- Use a simulated null distribution to estimate a p-value by counting “as extreme or more extreme” outcomes.
- Explain in words what the simulation is modeling (random chance under the null).
- Make a decision by comparing a simulation-based p-value to .
- Common mistakes:
- Forgetting “at least as extreme” when counting simulated outcomes.
- Using the wrong tail (counting extreme results in the wrong direction).
- Treating a simulation p-value as exact rather than an estimate.
Practical Issues: Robustness, Outliers, and Study Design Pitfalls
Robustness: why t procedures often work even when data aren’t perfect
Real quantitative data are rarely perfectly normal. Fortunately, t procedures are robust in many common situations, meaning they still perform reasonably well even when assumptions are not met exactly.
In general:
- With moderate skewness and no extreme outliers, t procedures work well for moderate to large .
- With strong skewness or outliers, you need larger samples for the CLT to make inference reliable.
- With small samples, outliers and strong skewness can seriously distort inference.
Outliers matter because both and are sensitive to extreme values. One outlier can inflate (widening intervals and reducing significance) or pull (shifting the estimate), or both.
What to do when you see skewness or outliers
On the AP Exam, you usually are not expected to “fix” data with complex methods, but you are expected to reason appropriately.
- If there is an outlier and the sample is small, express caution about using t procedures and explain why.
- If the sample is large, you can often proceed but should still acknowledge the unusual feature.
- If a problem explicitly asks whether conditions are met, mention the graph-based evidence.
Sometimes a transformation (like a logarithm) can help with skewed data, but the emphasis is typically on recognizing when t procedures are or are not appropriate.
Design problems inference cannot fix
Even perfect calculations cannot rescue a poor design. Major threats include:
- Bias in sampling (convenience samples, voluntary response): inference procedures do not correct bias.
- Nonresponse: if nonresponders differ systematically, results can be biased.
- Confounding in observational studies: even if a difference in means is statistically significant, it may be explained by lurking variables.
A useful mindset is: t procedures quantify random sampling variability, not systematic bias.
Statistical significance vs practical importance
A small p-value can occur because the effect is meaningful, or because the sample is large enough to detect even tiny differences. AP free-response questions sometimes ask for practical interpretation, so be ready to comment on effect size (like the estimated difference in means), not only whether it is statistically significant.
Example of a “trap” scenario to reason through
Suppose a very large sample produces a 99% confidence interval for of points on a 100-point test. This is statistically convincing evidence of a difference (0 is not included), but a difference of about half a point may be practically unimportant for real decisions.
Exam Focus
- Typical question patterns:
- Given a histogram/boxplot, decide whether a t procedure is appropriate and justify.
- Explain why a statistically significant result may still be practically unimportant.
- Identify whether a conclusion can be generalized or interpreted causally based on sampling/assignment.
- Common mistakes:
- Assuming “random sample” when the problem states a convenience sample.
- Ignoring an obvious outlier in a small-sample setting.
- Saying “significant” as if it automatically means “important.”
Type I Errors, Type II Errors, and Power
The ideas
When you run a significance test, there are two ways to be wrong.
- Type I error: rejecting a true null hypothesis.
- Type II error: failing to reject a false null hypothesis.
Power is the probability of rejecting a false null hypothesis (in other words, correctly detecting a real effect). If is the probability of a Type II error, then power is:
Example 7.7: Type II error and power (illustrated with a proportion)
A candidate claims to have the support of 70% of the people, but you believe the true figure is lower. You plan to gather an SRS and will reject the 70% claim if your sample shows 65% or less support. Suppose that, in reality, only 63% of the people support the candidate.
- The null model corresponds to .
- The decision rule is: reject if the sample proportion is 0.65 or less.
When will you fail to recognize that the null hypothesis is incorrect (a Type II error)? Precisely when the sample proportion is greater than 0.65. The probability of this happening (under the true model ) is .
When will you rightly conclude that the null hypothesis is incorrect? When the sample proportion is less than or equal to 0.65. The probability of that event (under ) is the power, .
This example is about proportions, but the logic of Type II error and power applies to tests for means as well.
Exam Focus
- Typical question patterns:
- Identify Type I vs Type II error in a context after a decision.
- Describe what “power” means in plain language.
- Explain (conceptually) when Type II errors happen under an alternative that is actually true.
- Common mistakes:
- Mixing up which error matches “reject” versus “fail to reject.”
- Describing Type I/II errors without referencing the context.
- Thinking power is a property of one dataset rather than a long-run probability under a specified alternative.