Last Minute AP Statistics Cheat Sheet (WITH FORMULAS)

What You Need to Know

This is the high-yield formula + procedure sheet you use when you’re trying to (1) pick the right method fast, (2) check conditions correctly, and (3) write the minimum-necessary but full-credit inference “story” (parameter → conditions → compute → conclude in context).

Big AP Stats idea: almost every FRQ is either describing data, probability/random variables, or inference (confidence interval or significance test). The fastest way to lose points is skipping conditions or failing to define the parameter.

Golden rule: Your hypotheses and conclusion must be about a population parameter (like pp, μ\mu, μ1μ2\mu_1-\mu_2, β\beta), not about sample statistics (like p^\hat p, xˉ\bar x).


Step-by-Step Breakdown

A. Picking the right inference procedure (fast decision tree)
  1. Is your response variable categorical (yes/no) or quantitative (number)?
    • Categorical → proportions, chi-square.
    • Quantitative → means, t-procedures, regression.
  2. How many groups/samples?
    • 1 sample → one-proportion zz or one-mean tt.
    • 2 independent samples → two-proportion zz or two-sample tt.
    • Matched pairs → one-sample tt on differences.
  3. Are you comparing distributions of categories across groups?
    • One categorical variable vs a claimed model → chi-square GOF.
    • Two categorical variables (relationship) → chi-square independence.
    • Several populations/treatments and one categorical response → chi-square homogeneity.
  4. Is it a relationship between two quantitative variables?
    • Use linear regression; inference about slope β\beta uses a tt test/interval with df=n2df=n-2.
B. Writing any inference solution (full-credit skeleton)
  1. Define the parameter (in context).
    • Example: pp = true proportion of all students at your school who…
  2. State hypotheses (test only).
    • H0:p=p0H_0: p=p_0, Ha:pp0H_a: p\ne p_0 (or <<, >>)
  3. Check conditions (name + verify with given info).
    • Random, 10% condition, Normal/Large Counts, etc.
  4. Compute the test statistic or interval (show formula + plug values).
  5. P-value OR critical value method (usually P-value on AP).
  6. Conclude in context at level α\alpha.
    • “Because p-value<αp\text{-value} < \alpha, reject H0H_0. There is convincing evidence that …”
C. Mini worked walkthrough (one-proportion zz test)

Prompt style: “Is there evidence the true proportion differs from 0.400.40?”

  1. Parameter: pp = true proportion of (population) who …
  2. Hypotheses: H0:p=0.40H_0: p=0.40, Ha:p0.40H_a: p\ne 0.40
  3. Conditions:
    • Random: stated random sample/assignment.
    • 10%: n0.1Nn \le 0.1N (if sampling without replacement).
    • Large counts: np010np_0\ge 10 and n(1p0)10n(1-p_0)\ge 10.
  4. Compute:
    • p^=xn\hat p = \dfrac{x}{n}
    • z=p^p0p0(1p0)nz=\dfrac{\hat p - p_0}{\sqrt{\dfrac{p_0(1-p_0)}{n}}}
  5. Get P-value from Normal.
  6. Conclude in context.

Key Formulas, Rules & Facts

A. Describing data (quick hits)
ToolFormulaWhen to useNotes
Standard scorez=xμσz=\dfrac{x-\mu}{\sigma} or z=xxˉsz=\dfrac{x-\bar x}{s}Compare to distribution center/spread“How many SDs from mean?”
Outlier rulebelow Q11.5(IQR)Q_1-1.5(IQR) or above Q3+1.5(IQR)Q_3+1.5(IQR)Boxplots/outliersNot “proof,” just a flag
Density/probabilityarea under curveContinuous modelsProbability = area
B. Linear transformations & combining variables
RuleFormulaNotes
Add constant aaμX+a=μX+a\mu_{X+a}=\mu_X+a, σX+a=σX\sigma_{X+a}=\sigma_XShifts center only
Multiply by bbμbX=bμX\mu_{bX}=b\mu_X, σbX=bσX\sigma_{bX}=|b|\sigma_XStretch/compress spread
Sum (any)μX+Y=μX+μY\mu_{X+Y}=\mu_X+\mu_YAlways true
Sum (independent)σX+Y2=σX2+σY2\sigma^2_{X+Y}=\sigma_X^2+\sigma_Y^2Variances add, not SDs
Difference (independent)σXY2=σX2+σY2\sigma^2_{X-Y}=\sigma_X^2+\sigma_Y^2Still add variances
C. Probability essentials
RuleFormulaUseNotes
ComplementP(Ac)=1P(A)P(A^c)=1-P(A)“At least one”Often fastest
Addition ruleP(AB)=P(A)+P(B)P(AB)P(A\cup B)=P(A)+P(B)-P(A\cap B)Two eventsIf disjoint, intersection =0=0
ConditionalP(AB)=P(AB)P(B)P(A\mid B)=\dfrac{P(A\cap B)}{P(B)}Given infoRestrict sample space
IndependenceP(AB)=P(A)P(B)P(A\cap B)=P(A)P(B)Check independenceEquivalent to P(AB)=P(A)P(A\mid B)=P(A)
BayesP(AB)=P(BA)P(A)P(B)P(A\mid B)=\dfrac{P(B\mid A)P(A)}{P(B)}Reverse conditionTree diagrams help
D. Discrete random variables (AP favorites)
ModelProbabilityMeanSDConditions/Notes
Binomial XBin(n,p)X\sim Bin(n,p)P(X=k)=(nk)pk(1p)nkP(X=k)=\binom{n}{k}p^k(1-p)^{n-k}μ=np\mu=npσ=np(1p)\sigma=\sqrt{np(1-p)}BINS: Binary, Independent, Number fixed, Same pp
Geometric XGeom(p)X\sim Geom(p)P(X=k)=(1p)k1pP(X=k)=(1-p)^{k-1}pμ=1p\mu=\dfrac{1}{p}σ=1pp2\sigma=\sqrt{\dfrac{1-p}{p^2}}Counts trials until first success
Expected valueμX=E(X)=xP(x)\mu_X=E(X)=\sum x\,P(x)Any discrete RVUse for “long-run average”
E. Normal + sampling distributions
IdeaFormulaWhen it appliesNotes
Normal modelXN(μ,σ)X\sim N(\mu,\sigma)Given approx NormalStandardize to use Normal CDF
Sample meanμxˉ=μ\mu_{\bar x}=\mu, σxˉ=σn\sigma_{\bar x}=\dfrac{\sigma}{\sqrt{n}}SRS; Normal pop or large nnCLT: large nn makes xˉ\bar x approx Normal
Sample proportionμp^=p\mu_{\hat p}=p, σp^=p(1p)n\sigma_{\hat p}=\sqrt{\dfrac{p(1-p)}{n}}Large countsFor inference, check large counts
Large counts (one prop)np10np\ge 10 and n(1p)10n(1-p)\ge 10Normal approx for p^\hat pFor tests use p0p_0 in check
Large counts (two prop)n1p110n_1p_1\ge 10, n1(1p1)10n_1(1-p_1)\ge 10, n2p210n_2p_2\ge 10, n2(1p2)10n_2(1-p_2)\ge 10Two-prop intervalsFor tests often use pooled p^\hat p
F. Confidence intervals (CI) and test statistics (most-used)
One proportion
TaskFormulaNotes
CI for ppp^±zp^(1p^)n\hat p \pm z^*\sqrt{\dfrac{\hat p(1-\hat p)}{n}}Use p^\hat p in SE
Test for ppz=p^p0p0(1p0)nz=\dfrac{\hat p-p_0}{\sqrt{\dfrac{p_0(1-p_0)}{n}}}Use p0p_0 in SE
Two proportions (independent)
TaskFormulaNotes
CI for p1p2p_1-p_2(p^1p^2)±zp^1(1p^1)n1+p^2(1p^2)n2(\hat p_1-\hat p_2) \pm z^*\sqrt{\dfrac{\hat p_1(1-\hat p_1)}{n_1}+\dfrac{\hat p_2(1-\hat p_2)}{n_2}}Don’t pool for CI
Test for p1p2p_1-p_2z=(p^1p^2)0p^(1p^)(1n1+1n2)z=\dfrac{(\hat p_1-\hat p_2)-0}{\sqrt{\hat p(1-\hat p)\left(\dfrac{1}{n_1}+\dfrac{1}{n_2}\right)}} where p^=x1+x2n1+n2\hat p=\dfrac{x_1+x_2}{n_1+n_2}Pool only in the test under H0:p1=p2H_0: p_1=p_2
One mean (quantitative)
TaskFormulaNotes
CI for μ\muxˉ±tsn\bar x \pm t^*\dfrac{s}{\sqrt{n}} with df=n1df=n-1Use when σ\sigma unknown (usual)
Test for μ\mut=xˉμ0s/nt=\dfrac{\bar x-\mu_0}{s/\sqrt{n}} with df=n1df=n-1Check approx Normal / no strong skew+outliers
Two means (independent samples)
TaskFormulaNotes
CI for μ1μ2\mu_1-\mu_2(xˉ1xˉ2)±ts12n1+s22n2(\bar x_1-\bar x_2) \pm t^*\sqrt{\dfrac{s_1^2}{n_1}+\dfrac{s_2^2}{n_2}}Calculator uses df approximation
Test for μ1μ2\mu_1-\mu_2t=(xˉ1xˉ2)0s12n1+s22n2t=\dfrac{(\bar x_1-\bar x_2)-0}{\sqrt{\dfrac{s_1^2}{n_1}+\dfrac{s_2^2}{n_2}}}Don’t pool SDs in AP Stats
Matched pairs (paired data)
  • Compute differences di=x1ix2id_i = x_{1i}-x_{2i}.
  • Then do one-sample tt on differences:
    • dˉ±tsdn\bar d \pm t^*\dfrac{s_d}{\sqrt{n}} and t=dˉμd,0sd/nt=\dfrac{\bar d-\mu_{d,0}}{s_d/\sqrt{n}} with df=n1df=n-1.
G. Chi-square procedures
ProcedureStatisticdfConditionsNotes
GOFχ2=(OE)2E\chi^2=\sum \dfrac{(O-E)^2}{E}k1k-1Random; expected counts typically 5\ge 5E=n×pmodelE=n\times p_{model}
Independence/Homogeneityχ2=(OE)2E\chi^2=\sum \dfrac{(O-E)^2}{E}(r1)(c1)(r-1)(c-1)Random; expected counts typically 5\ge 5E=(row total)(col total)nE=\dfrac{(row\ total)(col\ total)}{n}
H. Regression (least squares + inference)
QuantityFormulaNotes
LSRLy^=a+bx\hat y=a+bxPredict yy from xx
Slopeb=rsysxb=r\dfrac{s_y}{s_x}Sign matches rr
Intercepta=yˉbxˉa=\bar y-b\bar xLine goes through (xˉ,yˉ)\left(\bar x,\bar y\right)
Residuale=yy^e=y-\hat yPositive residual = point above line
Correlation1r1-1\le r\le 1No units; linear strength only
Coef. of determinationr2r^2% variability in yy explained by linear model with xx
Slope testt=b0SEbt=\dfrac{b-0}{SE_b}, df=n2df=n-2Test H0:β=0H_0: \beta=0
CI for slopeb±tSEbb\pm t^*SE_bInterpret change in mean response

Regression conditions (LINER): Linear pattern, Independent, Normal residuals, Equal variance, Random.

I. Inference vocabulary (quick definitions)
  • P-value: probability (assuming H0H_0 true) of getting a statistic as extreme or more extreme than observed.
  • Type I error: reject true H0H_0 (false positive). Probability =α=\alpha.
  • Type II error: fail to reject false H0H_0 (false negative). Probability =β=\beta.
  • Power: 1β1-\beta.

Examples & Applications

Example 1: Two-proportion zz interval (wording trap)

Situation: Compare vaccination rates in School A vs School B.

  • Parameter: pApBp_A-p_B = true difference in vaccination proportions.
  • Use CI:
    • (p^Ap^B)±zp^A(1p^A)nA+p^B(1p^B)nB\left(\hat p_A-\hat p_B\right) \pm z^*\sqrt{\dfrac{\hat p_A(1-\hat p_A)}{n_A}+\dfrac{\hat p_B(1-\hat p_B)}{n_B}}
      Key insight: If CI contains 00, a “difference” claim isn’t supported.
Example 2: Matched pairs vs two-sample (super common)

Situation: Same students take a pretest and posttest.

  • Don’t do two-sample tt.
  • Compute di=postpred_i=post-pre, then one-sample tt on μd\mu_d.
  • Test statistic: t=dˉ0sd/nt=\dfrac{\bar d-0}{s_d/\sqrt{n}}.
    Key insight: Pairing reduces variability; ignoring pairing can hide effects.
Example 3: Chi-square independence (interpretation)

Situation: Is seat location (front/middle/back) related to passing (yes/no)?

  • Parameter: whether the two categorical variables are independent in the population.
  • Expected count: E=(row total)(col total)nE=\dfrac{(row\ total)(col\ total)}{n}.
  • Statistic: χ2=(OE)2E\chi^2=\sum \dfrac{(O-E)^2}{E}, df=(r1)(c1)df=(r-1)(c-1).
    Key insight: A significant result says “associated,” not “causes.”
Example 4: Regression slope inference (what you conclude)

Situation: Predict exam score from hours studied.

  • Test H0:β=0H_0: \beta=0 vs Ha:β>0H_a: \beta>0.
  • Compute t=bSEbt=\dfrac{b}{SE_b} with df=n2df=n-2.
  • Conclusion in context: “There is convincing evidence of a positive linear relationship between hours studied and mean exam score.”
    Key insight: You’re making a claim about mean response changing with xx, not about individual predictions being perfect.

Common Mistakes & Traps

  1. Mistake: Hypotheses about p^\hat p or xˉ\bar x instead of pp or μ\mu.

    • Why wrong: sample stats are random; parameters are fixed truths.
    • Fix: define parameter first, then write H0H_0 and HaH_a about it.
  2. Mistake: Using tt vs zz incorrectly.

    • Why wrong: means with unknown σ\sigma require tt; proportions use zz.
    • Fix: quantitative → tt, categorical → zz.
  3. Mistake: Pooling in a two-proportion CI.

    • Why wrong: pooling assumes p1=p2p_1=p_2, which is exactly what you’re estimating in a CI.
    • Fix: Pool only for the hypothesis test of p1p2=0p_1-p_2=0.
  4. Mistake: Skipping or mis-checking large counts.

    • Why wrong: Normal approximation can fail badly with small expected successes/failures.
    • Fix: For one-prop tests use p0p_0; for intervals use p^\hat p.
  5. Mistake: Treating matched pairs as independent samples.

    • Why wrong: within-person pairing creates dependence; you must analyze differences.
    • Fix: If the same subject is measured twice (or paired units), do one-sample tt on dd.
  6. Mistake: Wrong chi-square df / expected counts.

    • Why wrong: df controls the reference distribution; wrong df → wrong P-value.
    • Fix: GOF df=k1df=k-1; two-way tables df=(r1)(c1)df=(r-1)(c-1); compute EE using row/col totals.
  7. Mistake: Regression conclusion implies causation.

    • Why wrong: observational studies can have confounding.
    • Fix: Only randomized experiments justify cause-and-effect.
  8. Mistake: “No significance” = “proved equal.”

    • Why wrong: failing to reject H0H_0 means insufficient evidence, not proof.
    • Fix: say “not enough evidence to conclude…”

Memory Aids & Quick Tricks

Trick / mnemonicHelps you rememberWhen to use
SOCSShape, Outliers, Center, SpreadDescribing distributions fast
BINSBinomial conditions: Binary, Independent, Number fixed, Same ppDecide binomial vs not
10% conditionIndependence when sampling without replacementAny sampling inference
PLANParameter, Label (hypotheses), Assumptions/conditions, Name test/intervalAny inference FRQ write-up
“Pool for test, not for CI”Two-proportion pooling ruleTwo-proportion inference
LINERLinear, Independent, Normal residuals, Equal variance, RandomRegression inference
“df = n-1, n-2, (r-1)(c-1)”df for one-sample tt, regression slope, chi-square tableDon’t lose df points
CUSSChi-square: Counts, Use expected, Sum (OE)2E\dfrac{(O-E)^2}{E}, Shape is right-skewChi-square setup + interpretation

Quick Review Checklist

  • [ ] You defined the parameter (with population + context).
  • [ ] Your H0H_0 and HaH_a are about the parameter, and direction matches the prompt.
  • [ ] You checked Random, 10%, and the correct Normal/Large Counts condition.
  • [ ] Proportions: zz procedures; Means: tt procedures; Paired: analyze differences.
  • [ ] Two-prop test uses pooled p^\hat p; two-prop CI does not.
  • [ ] You used the correct df: n1n-1 (one-sample/paired tt), n2n-2 (regression slope), (r1)(c1)(r-1)(c-1) (chi-square table).
  • [ ] Your conclusion is in context and matches the decision: reject vs fail to reject.
  • [ ] You didn’t claim causation unless it was a randomized experiment.

You’ve got this—run the checklist on every inference question and you’ll avoid the biggest point leaks.