Distribution of Sample Mean & Central Limit Theorem

Distribution of the Sample Mean

  • Why sample means vary
    • Each sample is only a subset → different observed values → different sample means (xˉ)(\bar{x}).
    • Leads to three guiding questions:
    1. What is the central/typical value of (xˉ)(\bar{x})?
    2. What is the variability (spread) of (xˉ)(\bar{x})?
    3. Does this variability follow a recognizable distributional pattern?
  • Symmetry expectation
    • For large nn the distribution of (xˉ)(\bar{x}) is roughly symmetric → (xˉ)(\bar{x}) falls below and above μ\mu about 50 % of the time.
    • For small nn symmetry may fail, but most (xˉ)(\bar{x}) values still cluster near μ\mu.
  • Theoretical guarantee: μxˉ=μ\mu_{\bar{x}} = \mu (mean of all sample means equals the population mean).

Unbiasedness & the Bull’s-Eye Metaphor

  • Unbiased estimator: Expected value of estimator equals the parameter.
    • E[xˉ]=μE[\bar{x}] = \muxˉ\bar{x} is unbiased for μ\mu.
    • Other unbiased estimators mentioned:
    • Sample proportion p^\hat{p} for population proportion pp.
    • Sample variance s2s^2 for population variance σ2\sigma^2.
  • Bull’s-eye visual
    • Target center = true parameter.
    • Four scenarios:
    1. Wide scatter but centered → unbiased, high variance.
    2. Tight cluster, centered → unbiased, low variance (ideal).
    3. Wide scatter, shifted center → biased, high variance.
    4. Tight cluster, shifted center → biased, low variance.

Variability of the Sample Mean (Standard Error)

  • Infinite population (or sampling with replacement):
    • σxˉ=σn\sigma_{\bar{x}} = \dfrac{\sigma}{\sqrt{n}}
  • Finite population (sampling without replacement):
    • σxˉ=NnN1×σn\sigma_{\bar{x}} = \sqrt{\dfrac{N - n}{N - 1}} \times \dfrac{\sigma}{\sqrt{n}}
    • NN = population size, nn = sample size.
  • Key implication: Increasing nnσxˉ\sigma_{\bar{x}} → estimates become more precise.

Numerical Visualization Example

  • Population parameters: μ=43660\mu = 43\,660, σ=2500\sigma = 2\,500.
  • Three normal curves for xˉ\bar{x}:
    • n=25n = 25 → widest, flattest curve (largest standard error).
    • n=100n = 100 → intermediate width/height.
    • n=200n = 200 → tallest, narrowest curve (smallest standard error).

Desirable Properties of xˉ\bar{x}

  1. Unbiasedness: μxˉ=μ\mu_{\bar{x}} = \mu.
  2. Efficiency with nn: Larger nn ⇒ smaller σxˉ\sigma_{\bar{x}}.

Central Limit Theorem (CLT)

  • Often called the fundamental theorem of statistics.
  • Statement (for "sufficiently large" n30n \ge 30):
    1. Distribution of xˉ\bar{x} is approximately normal, regardless of the population’s shape.
    2. Center: μxˉ=μ\mu_{\bar{x}} = \mu.
    3. Spread: σxˉ=σn\sigma_{\bar{x}} = \dfrac{\sigma}{\sqrt{n}}.
  • Special case: If the population itself is normal, all three properties hold for any nn (even n < 30).

Visual Demonstrations of the CLT

  • Four population shapes examined:
    1. Bimodal (two humps).
    2. Uniform (all values equally likely).
    3. Exponential (right-skewed, higher likelihood near zero).
    4. Normal.
  • In every scenario, the sampling distribution of xˉ\bar{x} becomes (approximately) normal as n30n \to 30 or larger.
    • For the inherently normal population, normality of xˉ\bar{x} holds even at small nn.

Error of Estimation (Sampling Error)

  • Defined as difference between a sample statistic and the corresponding population parameter.
    • For means: Error=xˉμ\text{Error} = \bar{x} - \mu.
    • Also called sampling error or estimation error.
  • Arises purely because we observe a sample rather than the full population.
  • Can be reduced by:
    • Increasing nn (reduces standard error).
    • Using better sampling designs to avoid bias.

These notes cover the distribution, properties, and practical implications of the sample mean, capped by the Central Limit Theorem’s assurance that large-sample means behave normally and predictably.