Measures of Dispersion — Section 3.2 Notes
Range
Definition: The range measures the spread between the smallest and largest values in a data set.
Formula: R = \text{Largest Value} - \text{Smallest Value}
Example: Data set 125, 175, 200, 225, 250
Largest = 250, Smallest = 125
Range = 250 − 125 = 125
Pros and cons:
Pros: Very simple and quick to compute
Cons: Only uses two data values (extremes)
Susceptible to extreme values; NOT resistant to outliers
Standard deviation
Concept:
Based on deviations from the mean
Deviations sum to zero by definition
Describes the "typical" deviation from the mean
Other options (MAD, variance) exist but are less intuitive for describing spread around the mean
Population standard deviation
Formula: \sigma = \sqrt{\frac{1}{N} \sum{i=1}^{N} (xi - \mu)^2}
Interpretation: Average squared deviation from the mean, then take the square root
Sample standard deviation
Formula: s = \sqrt{\frac{1}{n-1} \sum{i=1}^{n} (xi - \bar{x})^2}
Uses n−1 in the denominator to make the estimator unbiased for the population SD
Hand calculation illustration (population, N=5): data 1, 2, 3, 4, 5
Mean: \mu = \frac{1+2+3+4+5}{5} = 3
Deviations: (-2,-1,0,1,2)
Squared deviations: (4,1,0,1,4)
Sum: \sum (x_i-\mu)^2 = 10
Population variance: \sigma^2 = \frac{10}{5} = 2
Population SD: \sigma = \sqrt{2} \approx 1.4142
Example with a sample (same data, n=5):
Sum of squared deviations: 10
Sample variance: s^2 = \frac{10}{n-1} = \frac{10}{4} = 2.5
Sample SD: s = \sqrt{2.5} \approx 1.5811
Calculation tables: Conceptual tool to organize the steps (x, x−μ, (x−μ)²) for both population and sample calculations
A few notes about standard deviation
Rounding:
Be careful with rounding intermediate calculations; small changes can affect the final SD value
Interpretation:
Large deviations from the mean lead to a larger SD; small deviations lead to a smaller SD
Resistance:
Standard deviation is NOT resistant to outliers; a few extreme values can substantially increase SD
n−1 denominator (why):
The use of n−1 makes the estimator of the population variance unbiased when sampling
Comparing standard deviations
When to compare:
If two samples/populations share the same units and are on the same scale, SDs can be directly compared
When units differ:
Use the coefficient of variation (CV), a unitless measure of relative spread
Coefficient of variation (CV):
Formula: \text{CV} = \frac{\sigma}{|\mu|}
Interpretation: SD expressed as a fraction of the mean; enables comparison across different units/scales
Variance
Relationship to SD:
Variance is the square of the standard deviation
Population variance: \sigma^2 = \frac{1}{N} \sum{i=1}^{N} (xi - \mu)^2
Sample variance: s^2 = \frac{1}{n-1} \sum{i=1}^{n} (xi - \bar{x})^2
Interpretation:
In the hand-calculation context, variance is the quantity inside the square root
Units are squared, which can make interpretation less intuitive
Role in inferential statistics:
Variance is foundational for many inferential procedures, including hypothesis testing and confidence intervals
Notes about StatCrunch
Mean:
The mean formula is the same for samples and populations; StatCrunch treats the mean as the same quantity in either case
Standard deviation in StatCrunch:
The default “standard deviation” in StatCrunch is the sample SD
To obtain the population SD, choose “unadj. standard deviation”
Practical takeaway for exams:
Know how to compute by hand (sum of squared deviations, divide by N or n−1, take the square root)
The empirical rule (68-95-99.7 rule)
Typically described for bell-shaped (normal) distributions:
About 68% of data lie within 1 standard deviation of the mean
About 95% lie within 2 standard deviations
About 99.7% lie within 3 standard deviations
The empirical rule also aligns with the idea of a normal distribution where most data cluster near the mean
Visual shorthand:
1 SD: 68%
2 SD: 95%
3 SD: 99.7%
Special tail details (for normal):
About 0.15% lie beyond 3 SD on each tail (total ~0.3% beyond |Z|>3)
About 2.35% lie beyond 2 SD on each tail (total ~4.7% beyond |Z|>2)
Example: Bell-shaped distribution with mean 150 and standard deviation 15
68% of observations lie between which values?
Between 150 - 15 = 135 and 150 + 15 = 165
Interval: [135, 165]
95% of observations lie between which values?
Between 150 - 2\times 15 = 120 and 150 + 2\times 15 = 180
Interval: [120, 180]
What percentage lie between 105 and 195?
This is within 3 standard deviations: 150 \pm 3\cdot 15 = 105, 195
Percentage: approximately 99.7%
Chebyshev’s inequality
Scope:
Applies to any distribution, not just bell-shaped
Statement:
For any distribution with mean (\mu) and standard deviation (\sigma), at least \left(1 - \frac{1}{k^2}\right) \times 100\% of observations lie within (k) standard deviations of the mean
Key takeaway:
Guarantees a minimum proportion of observations within (k\sigma) of the mean, regardless of distribution shape
Example: Chebyshev (mean = 150, SD = 15)
Question: At least what percentage lie between 120 and 180?
Here, 120 and 180 are within 2 standard deviations of the mean ((k=2))
Minimum percentage: \left(1 - \frac{1}{2^2}\right) \times 100\% = \left(1 - \frac{1}{4}\right) \times 100\% = 75\%
Empirical rule vs Chebyshev: a quick comparison
Distribution requirements:
Empirical rule assumes a bell-shaped (normal) distribution
Chebyshev’s inequality places no assumption on shape; it applies to any distribution
Nature of the statement:
Empirical rule gives approximate percentages for normal data
Chebyshev provides a guaranteed minimum ("at least"), not an exact proportion
Practical interpretation:
If you know the data are roughly normal, use the empirical rule for quick estimates
If the distribution shape is unknown or non-normal, use Chebyshev to obtain a conservative bound
Quick recap and study-ready takeaways
Range: simplest spread measure; sensitive to outliers
Standard deviation: core measure of spread around the mean; population SD uses N in the denominator, sample SD uses n−1
Variance: the squared SD; easier to manipulate in algebraic/statistical derivations; units are squared
Coefficient of variation: unitless way to compare spread across different means/units
The empirical rule: best for normal distributions; gives quick spread estimates
Chebyshev’s inequality: universal spread bound; useful when distribution is unknown
When using software (e.g., StatCrunch): know which SD you’re getting by default and how to obtain the population version if needed