Chapter 15- Sampling Distributions
Key Concepts in Sampling Distributions
Parameters and Statistics:
Parameters are numerical values that describe a population (e.g., population mean (μ), population standard deviation (σ)).
Statistics are numerical values that can be computed from sample data and are often used to estimate parameters (e.g., sample mean (x̄), sample standard deviation (s)).
It's crucial to differentiate between these when analyzing data to avoid confusion in drawing conclusions.
Statistical Estimation:
Statistical inference uses sample data to infer about the population.
Different samples yield different statistics; thus, understanding the sampling distribution is vital for making valid inferences.
Law of Large Numbers:
This principle states that as sample size increases, the sample mean (x̄) will get closer to the population mean (μ).
It guarantees that larger samples provide a more accurate estimation of the population parameters.
Sampling Distributions:
The distribution of values taken by a statistic in all possible samples of the same size from a population is called a sampling distribution.
Notably, the sampling distribution of the sample mean is often less spread out than the population distribution.
Population vs. Sampling Distributions:
Population Distribution: Describes how individual values vary in the entire population.
Sampling Distribution: Describes how sample means vary across repeated samples, typically tighter and centered around the population mean.
Sampling Distribution Characteristics:
The mean of the sampling distribution of a sample mean is μ, and the standard deviation is σ/√n, where n is the sample size.
This means that as the sample size increases, the variability of the sample means decreases.
Shapes of Sampling Distributions:
If the population distribution is Normal, the sampling distribution is also Normal.
If the population distribution is not Normal, the Central Limit Theorem (CLT) applies: as n increases, the sampling distribution of the sample mean approaches a Normal distribution regardless of the shape of the population distribution.
Central Limit Theorem (CLT):
States that for large sample sizes (usually n ≥ 30), the sampling distribution of the sample mean will be approximately Normal, which allows for the use of Normal probability models even if the population distribution is not.
Statistical Significance:
An observed effect that rarely occurs due to chance is considered statistically significant.
The sampling distribution allows us to determine the probability of observing a statistic in future samples and identify whether an observed statistic is significant.
Examples:
In a case involving technician times for AC maintenance, the sampling distribution of the mean time can be analyzed using CLT to determine the likelihood of completing work within a specific budget.
If the distribution is approximately Normal, probability calculations can predict outcomes based on the sample mean.
Summary of Important Formulas:
Mean of Sampling Distribution: μ (mean of population)
Standard Deviation of Sampling Distribution: σ/√n
Sampling Distribution (as per CLT): If individuals are from N(μ, σ), then sample mean from n will be N(μ, σ/√n).
Understanding these concepts is crucial for conducting proper statistical analysis and for making valid inferences based on sample data.