Last Minute AP Statistics Cheat Sheet (WITH FORMULAS)
What You Need to Know
This is the high-yield formula + procedure sheet you use when you’re trying to (1) pick the right method fast, (2) check conditions correctly, and (3) write the minimum-necessary but full-credit inference “story” (parameter → conditions → compute → conclude in context).
Big AP Stats idea: almost every FRQ is either describing data, probability/random variables, or inference (confidence interval or significance test). The fastest way to lose points is skipping conditions or failing to define the parameter.
Golden rule: Your hypotheses and conclusion must be about a population parameter (like , , , ), not about sample statistics (like , ).
Step-by-Step Breakdown
A. Picking the right inference procedure (fast decision tree)
- Is your response variable categorical (yes/no) or quantitative (number)?
- Categorical → proportions, chi-square.
- Quantitative → means, t-procedures, regression.
- How many groups/samples?
- 1 sample → one-proportion or one-mean .
- 2 independent samples → two-proportion or two-sample .
- Matched pairs → one-sample on differences.
- Are you comparing distributions of categories across groups?
- One categorical variable vs a claimed model → chi-square GOF.
- Two categorical variables (relationship) → chi-square independence.
- Several populations/treatments and one categorical response → chi-square homogeneity.
- Is it a relationship between two quantitative variables?
- Use linear regression; inference about slope uses a test/interval with .
B. Writing any inference solution (full-credit skeleton)
- Define the parameter (in context).
- Example: = true proportion of all students at your school who…
- State hypotheses (test only).
- , (or , )
- Check conditions (name + verify with given info).
- Random, 10% condition, Normal/Large Counts, etc.
- Compute the test statistic or interval (show formula + plug values).
- P-value OR critical value method (usually P-value on AP).
- Conclude in context at level .
- “Because , reject . There is convincing evidence that …”
C. Mini worked walkthrough (one-proportion test)
Prompt style: “Is there evidence the true proportion differs from ?”
- Parameter: = true proportion of (population) who …
- Hypotheses: ,
- Conditions:
- Random: stated random sample/assignment.
- 10%: (if sampling without replacement).
- Large counts: and .
- Compute:
- Get P-value from Normal.
- Conclude in context.
Key Formulas, Rules & Facts
A. Describing data (quick hits)
| Tool | Formula | When to use | Notes |
|---|---|---|---|
| Standard score | or | Compare to distribution center/spread | “How many SDs from mean?” |
| Outlier rule | below or above | Boxplots/outliers | Not “proof,” just a flag |
| Density/probability | area under curve | Continuous models | Probability = area |
B. Linear transformations & combining variables
| Rule | Formula | Notes |
|---|---|---|
| Add constant | , | Shifts center only |
| Multiply by | , | Stretch/compress spread |
| Sum (any) | Always true | |
| Sum (independent) | Variances add, not SDs | |
| Difference (independent) | Still add variances |
C. Probability essentials
| Rule | Formula | Use | Notes |
|---|---|---|---|
| Complement | “At least one” | Often fastest | |
| Addition rule | Two events | If disjoint, intersection | |
| Conditional | Given info | Restrict sample space | |
| Independence | Check independence | Equivalent to | |
| Bayes | Reverse condition | Tree diagrams help |
D. Discrete random variables (AP favorites)
| Model | Probability | Mean | SD | Conditions/Notes |
|---|---|---|---|---|
| Binomial | BINS: Binary, Independent, Number fixed, Same | |||
| Geometric | Counts trials until first success | |||
| Expected value | Any discrete RV | Use for “long-run average” |
E. Normal + sampling distributions
| Idea | Formula | When it applies | Notes |
|---|---|---|---|
| Normal model | Given approx Normal | Standardize to use Normal CDF | |
| Sample mean | , | SRS; Normal pop or large | CLT: large makes approx Normal |
| Sample proportion | , | Large counts | For inference, check large counts |
| Large counts (one prop) | and | Normal approx for | For tests use in check |
| Large counts (two prop) | , , , | Two-prop intervals | For tests often use pooled |
F. Confidence intervals (CI) and test statistics (most-used)
One proportion
| Task | Formula | Notes |
|---|---|---|
| CI for | Use in SE | |
| Test for | Use in SE |
Two proportions (independent)
| Task | Formula | Notes |
|---|---|---|
| CI for | Don’t pool for CI | |
| Test for | where | Pool only in the test under |
One mean (quantitative)
| Task | Formula | Notes |
|---|---|---|
| CI for | with | Use when unknown (usual) |
| Test for | with | Check approx Normal / no strong skew+outliers |
Two means (independent samples)
| Task | Formula | Notes |
|---|---|---|
| CI for | Calculator uses df approximation | |
| Test for | Don’t pool SDs in AP Stats |
Matched pairs (paired data)
- Compute differences .
- Then do one-sample on differences:
- and with .
G. Chi-square procedures
| Procedure | Statistic | df | Conditions | Notes |
|---|---|---|---|---|
| GOF | Random; expected counts typically | |||
| Independence/Homogeneity | Random; expected counts typically |
H. Regression (least squares + inference)
| Quantity | Formula | Notes |
|---|---|---|
| LSRL | Predict from | |
| Slope | Sign matches | |
| Intercept | Line goes through | |
| Residual | Positive residual = point above line | |
| Correlation | No units; linear strength only | |
| Coef. of determination | % variability in explained by linear model with | |
| Slope test | , | Test |
| CI for slope | Interpret change in mean response |
Regression conditions (LINER): Linear pattern, Independent, Normal residuals, Equal variance, Random.
I. Inference vocabulary (quick definitions)
- P-value: probability (assuming true) of getting a statistic as extreme or more extreme than observed.
- Type I error: reject true (false positive). Probability .
- Type II error: fail to reject false (false negative). Probability .
- Power: .
Examples & Applications
Example 1: Two-proportion interval (wording trap)
Situation: Compare vaccination rates in School A vs School B.
- Parameter: = true difference in vaccination proportions.
- Use CI:
Key insight: If CI contains , a “difference” claim isn’t supported.
Example 2: Matched pairs vs two-sample (super common)
Situation: Same students take a pretest and posttest.
- Don’t do two-sample .
- Compute , then one-sample on .
- Test statistic: .
Key insight: Pairing reduces variability; ignoring pairing can hide effects.
Example 3: Chi-square independence (interpretation)
Situation: Is seat location (front/middle/back) related to passing (yes/no)?
- Parameter: whether the two categorical variables are independent in the population.
- Expected count: .
- Statistic: , .
Key insight: A significant result says “associated,” not “causes.”
Example 4: Regression slope inference (what you conclude)
Situation: Predict exam score from hours studied.
- Test vs .
- Compute with .
- Conclusion in context: “There is convincing evidence of a positive linear relationship between hours studied and mean exam score.”
Key insight: You’re making a claim about mean response changing with , not about individual predictions being perfect.
Common Mistakes & Traps
Mistake: Hypotheses about or instead of or .
- Why wrong: sample stats are random; parameters are fixed truths.
- Fix: define parameter first, then write and about it.
Mistake: Using vs incorrectly.
- Why wrong: means with unknown require ; proportions use .
- Fix: quantitative → , categorical → .
Mistake: Pooling in a two-proportion CI.
- Why wrong: pooling assumes , which is exactly what you’re estimating in a CI.
- Fix: Pool only for the hypothesis test of .
Mistake: Skipping or mis-checking large counts.
- Why wrong: Normal approximation can fail badly with small expected successes/failures.
- Fix: For one-prop tests use ; for intervals use .
Mistake: Treating matched pairs as independent samples.
- Why wrong: within-person pairing creates dependence; you must analyze differences.
- Fix: If the same subject is measured twice (or paired units), do one-sample on .
Mistake: Wrong chi-square df / expected counts.
- Why wrong: df controls the reference distribution; wrong df → wrong P-value.
- Fix: GOF ; two-way tables ; compute using row/col totals.
Mistake: Regression conclusion implies causation.
- Why wrong: observational studies can have confounding.
- Fix: Only randomized experiments justify cause-and-effect.
Mistake: “No significance” = “proved equal.”
- Why wrong: failing to reject means insufficient evidence, not proof.
- Fix: say “not enough evidence to conclude…”
Memory Aids & Quick Tricks
| Trick / mnemonic | Helps you remember | When to use |
|---|---|---|
| SOCS | Shape, Outliers, Center, Spread | Describing distributions fast |
| BINS | Binomial conditions: Binary, Independent, Number fixed, Same | Decide binomial vs not |
| 10% condition | Independence when sampling without replacement | Any sampling inference |
| PLAN | Parameter, Label (hypotheses), Assumptions/conditions, Name test/interval | Any inference FRQ write-up |
| “Pool for test, not for CI” | Two-proportion pooling rule | Two-proportion inference |
| LINER | Linear, Independent, Normal residuals, Equal variance, Random | Regression inference |
| “df = n-1, n-2, (r-1)(c-1)” | df for one-sample , regression slope, chi-square table | Don’t lose df points |
| CUSS | Chi-square: Counts, Use expected, Sum , Shape is right-skew | Chi-square setup + interpretation |
Quick Review Checklist
- [ ] You defined the parameter (with population + context).
- [ ] Your and are about the parameter, and direction matches the prompt.
- [ ] You checked Random, 10%, and the correct Normal/Large Counts condition.
- [ ] Proportions: procedures; Means: procedures; Paired: analyze differences.
- [ ] Two-prop test uses pooled ; two-prop CI does not.
- [ ] You used the correct df: (one-sample/paired ), (regression slope), (chi-square table).
- [ ] Your conclusion is in context and matches the decision: reject vs fail to reject.
- [ ] You didn’t claim causation unless it was a randomized experiment.
You’ve got this—run the checklist on every inference question and you’ll avoid the biggest point leaks.