Probability Bounds and Central Limit Theorems Study Notes
Probability and Stochastic Processes
Unit III: Probability Bounds and Central Limit Theorems
Department of Mathematics, SRM Institute of Science and Technology, Kattankulathur-603203
Introduction
This unit discusses probability bounds, which are inequalities applicable in various scenarios.
Purpose of using inequalities:
Insufficient information to calculate a desired quantity (e.g., probability of an event, expected value of a random variable).
Complicated problems where exact calculations are difficult.
Provide general results applicable to a wide range of problems.
In this section, µ and σ² represent the mean and variance of the random variable under consideration.
Markov’s Inequality
General Theme
Focus on upper bounding tail probabilities, i.e., probabilities of the forms
.
Definition
Markov’s Inequality:
Let X be a non-negative random variable.
Then, for any a > 0 :
.
Proof of Markov’s Inequality
Discrete Random Variable
Assume X is a discrete random variable.
The expected value is defined as:
This can be split into two parts:
\sum_{x<a} xP(X=x) + \sum_{x\geq a} xP(X=x).
Since X is non-negative,
Rearranging yields:
Continuous Random Variable
Assume X is a continuous random variable.
The expected value is defined as:
The inequality is similarly derived:
Examples of Markov’s Inequality
Example 1
Let .
Apply Markov's inequality to find an upper bound on , where p < \alpha < 1 .
Solution:
.
With and , it follows that
Example 2
Let .
Determine an upper bound for , where a > 0 :
The p.d.f. of is given by:
..
Thus, using Markov's inequality:
The actual value is given by:
Chebyshev’s Inequality
Definition
Chebyshev’s Inequality:
If X is a random variable with
Mean and
Variance ,
Then:
,or:
P(|X − \mu| < a) \geq 1 - \frac{\sigma^2}{a^2} for any a > 0 .
Explanation
Chebyshev’s Inequality states that the probability of the difference between X and is limited by the variance of X.
Alternative Forms
If we let where k > 0 then:
Conversely:
P(|X − \mu| < kσ) \geq 1 - \frac{1}{k^2}.
Example 3
Show by Chebyshev’s inequality that for 2000 throws of a coin, the probability that the number of heads lies between 900 and 1100 is at least :
Let X denote the number of heads in 2000 throws of a coin:
,
.
Therefore,
P(900 < X < 1100) = P(-100 < X − 1000 < 100) = 1 - P(|X − 1000| \geq 100).By Chebyshev's inequality:
Thus,
P(900 < X < 1100) \geq 1 - \frac{1}{20} = \frac{19}{20}.
Example 4
For a random variable X with density function: .
Show that Chebyshev's inequality provides:
, while the actual probability is :
Calculate the mean:
.
Calculate the second moment:
Therefore,
.
Using Chebyshev’s inequality for any a > 0 :
Taking :
Calculating the actual probability:
P(|X - 1| \geq 2) = 1 - P(-1 < X < 3) = 1 - \int_{0}^{3} e^{-x}dx = 1 - (1 - e^{-3}) = e^{-3}.
Additional Concepts
Laws of Large Numbers
Convergence in Probability
A sequence of random variables converges in probability to , if:
For every \epsilon > 0 ,
\lim_{n \to \infty} P(|X_n - \alpha| < \epsilon) = 1OR
.
Weak Law of Large Numbers (WLLN)
Let be a sequence of random variables with respective means .
Let
Then:
provided:
, where B_n = Var(X_1 + X_2 + … + X_n) < \infty .
Conditions for WLLN
To verify WLLN holds:
(i) exists for all .
(ii) exists.(iii) .
Example 9
Check if WLLN holds for the sequence defined by
Calculate:
,
Thus,
B_n = Var(X_1 + … + X_n) = n < \infty .
Therefore,
.
Central Limit Theorem (CLT)
Definition
The Central Limit Theorem states:
The normal distribution is the limiting distribution of the sum of independent random variables with finite variance.
For , let be independent, identically distributed random variables with finite mean and variance , then:
As :
(a)
P\left( \frac{S_n - n\mu}{\sqrt{n}\sigma} ≤ x \right) \rightarrow \frac{1}{\sqrt{2C0}} \int_{-\infty}^{x} e^{-\frac{t^2}{2}} dt, \text{ for all } x \in \mathbb{R}.
(b)In summary:
For large n:
,
.
Remark on CLT Usage
For iid random variables, if n is large enough,
follows normal distribution with mean and variance ;
follows normal distribution with mean and variance .
Examples of CLT
Example 12
For a coin tossed 200 times, find the probability of number of heads between 80 and 120:
Let = number of heads. Then .
Thus,
Mean:
Variance:
.
Apply CLT:
Find
follows standard normal distribution.
Therefore, the probability is:
P(80 < X < 120) = P\left(-2.82 < Z < 2.82\right) = 0.9952 .
Example 13
Given 75 Poisson variates with parameter 2, estimate P(120 < S_n < 160) :
For ,
, .
By CLT:
.
The z-score transition leads to
P(120 < S_n < 160) = P(-2.4495 ≤ Z ≤ 0.8165) = 0.7868.
Example 14
For IID random variables with
estimate for :
Therefore,
.
Proceed similarly as before for CDF calculations:
.
Example 15
For the lifetime of bulbs with
, , find:
Probability that average lifetime of 60 bulbs exceeds 1250 hrs:
By CLT, .
Required probability:
P(\bar{X} > 1250) = P\left( Z > \frac{1250 - 1200}{250/\sqrt{60}} \right) = P(Z > 1.55) = 0.0606.
Strong Law of Large Numbers
For independent, identically distributed random variables with a finite expected value $ E(X_i) = \mu < \infty
It holds with probability 1 that as ,
i.e. .
One-sided Chebyshev’s Inequality
Definition
For random variable X with mean 0 and finite variance $ σ^2
For any c > 0 :
Example 16
Given items produced with mean 100 and variance 400, find :
Using one-sided Chebyshev’s:
From Markov’s inequality, .
Cauchy-Schwarz Inequality
Definition
If X and Y are two random variables such that , , and exist, then:
.
Equality holds if and only if:
, or for some constant a.
Example 17
Utilize Cauchy-Schwarz to prove:
- , where ρ = correlation coefficient.
Integrable Random Variables
Definition
A random variable X is integrable if its expected value exists, i.e., E(X) < \infty .
Convex Function
Definition
A twice-differentiable function is convex if the second derivative exists and is non-negative within its domain, i.e., .
Jensen’s Inequality
Jensen’s Inequality:
For integrable random variable X and a convex function ,
Then:
Example 18
Let X have mean 10, find:
Let
The second derivative $g''(x) > 0$ for x > -1 implies convexity:
Therefore:
Moment Generating Function
Definition
The moment generating function (M.G.F.) of a random variable X is defined as:
Chernoff Bounds
Definition
For any random variable X and real number a,
We can express:
P(X \geq a) = P(e^{tX} \geq e^{ta}), \text{ for } t > 0;
P(X \leq a) = P(e^{tX} \geq e^{ta}), \text{ for } t < 0.
Apply Markov’s inequality:
For t > 0:
Similarly for t < 0 :
Conclusion
Thus, we establish:
P(X \geq a) \leq ext{Min}{t>0} \left{ e^{-ta} M_X(t) \right}, ext{ for all } t > 0 P(X \leq a) \leq ext{Min}{t<0} \left{ e^{-ta} M_X(t) \right}, ext{ for all } t < 0.
Example 19
For a standard normal variable Z with its M.G.F.:
Estimate using Chernoff bounds:
P(Z ≥ 2) ≤ ext{Min} \left{ e^{-2t} M_Z(t) \right} .
Minimize the function:
Take derivative to identify critical points.
Get best upper bound
Comparing with one-sided Chebyshev’s:
Example 20
For a Poisson variate with parameter :
Calculate using Chernoff bounds:
P(X ≥ 3) ≤ ext{Min} \left{ e^{-3t} M_X(t) \right}.Establish CDF and compare with Markov's Inequality.
Construct the minimizing function and identify critical points accordingly.
Thank You
Session concluded with acknowledgments from
Department of Mathematics, SRM Institute of Science and Technology, Kattankulathur-603203.
Example 5
Let .
Find an upper bound for :
Using Markov's inequality:
P(X ext{ } ext{is } ext{ at } ext{ least } 20) iggr ext{ } \ ext{ }iggr ext{ } ext{ }iggr ext{ } ext{ } ext{ }iggr ext{ }iggr ext{ } \ ext{ } ext{ }iggr ext{ }iggr ext{ } \ ext{ } ext{ } iggr ext{ }iggr ext{ } \ ext{ } = rac{E(X)}{20} = rac{10}{20} = rac{1}{2}.
Example 6
Let Y ext{ be a random variable that takes values } 0 ext{ and } 3 P(Y = 3) E(Y) = 0.5 imes 0 + 0.5 imes 3 = 1.5. P(Y ext{ } ext{is } ext{ at } for ext{ least } 3) = rac{1.5}{3} = 0.5. Z ext{ follows a Uniform distribution on } [0, 10] E(Z) = rac{0 + 10}{2} = 5 P(Z ext{ > } 8) P(Z ext{ > } 8) iggr ext{ }iggr ext{ } = ext{ } ext{ at } \ ext{ } iggr ext{ }iggr ext{ } = rac{5}{8}. W E(W) = 4 P(W > 5) P(W > 5) iggr ext{ } = rac{E(W)}{5} = rac{4}{5}. $$