Probability Theory

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/73

flashcard set

Earn XP

Description and Tags

UW Madison Stat Qual Prep Option B

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

74 Terms

1
New cards

Theorem 1: Let X \subset R^d. Suppose that f: X \rightarrow R is non-negative and integrable. Then, …

There is a distribution P over X such that if x \sim P, for every A \subset X, we have \mathbb{P}(X \in A) = \int_{A} p(\textbf{x})d\textbf{x} where p(\textbf{x}) = \frac{f(\textbf{x})}{\int_Xf(\textbf{x})d\textbf{x}}

2
New cards

Bernoulli Density

p(x) = p^x(1-p)^{1-x} for x \in {0,1}

3
New cards

Bernoulli Expected Value

p

4
New cards

Bernoulli Variance

p(1-p)

5
New cards

Poisson Distribution Density

p(x) = \frac{\lambda^xe^{-\lambda}}{x!} for x \in \{0,1,\cdots\}

6
New cards

Poisson Expected Value

\lambda

7
New cards

Poisson Variance

\lambda

8
New cards

Gamma Function

\Gamma(x) = \int_{0}^{\infty}t^{x-1}e^{-t}dt

9
New cards

Gamma Function Properties

\Gamma(0) = \Gamma(1) = 1, \Gamma(x+1) = x\Gamma(x), \Gamma(n) = (n-1)! , \Gamma(1/2) = \sqrt{\pi}

10
New cards

Gamma Distribution Density

s,r > 0, p(x) = \frac{r^s}{\Gamma(s)}x^{s-1}e^{-rx} for x>0

11
New cards

Gamma Distribution Expected Value

s/r

12
New cards

Gamma Distribution Variance

s/r²

13
New cards

Beta Function

B(a,b) = \int_0^1 t^{a-1}(1-t)^{b-1}dt for a,b>0

14
New cards

Alternative Beta Function

B(a,b) = \frac{\Gamma(a) \Gamma(b)}{\Gamma(a+b)}

15
New cards

Standard Normal Density

p(x) = (2\pi)^{-1/2}e^{-x²/2}

16
New cards

Standard Normal Expected Value

0

17
New cards

Standard Normal Variance

1

18
New cards

Basis of Monte Carlo Approximations

Approximation of probability based on random samples

19
New cards

Relation between marginal and joint distribution

Can determine marginal distributions exactly and uniquely from the joint density, but cannotdetermine unique and exact joint density from the marginals

20
New cards

Marginal Density

Integrate out all variables that we are not interested in

21
New cards

Negative Binomial Density

p(x) = {{x-1}\choose{r-1}} p^r(1-p)^{x-r} , x = # of trials until rth success with probability of success p in each trial

22
New cards

Negative Binomial Expected Value

r/p

23
New cards

Negative Binomial Variance

r(1-p)/p²

24
New cards

Negative Binomial MGF

\left [ \frac{pe^t}{1-(1-p)e^t} \right ]^r

25
New cards

Poisson MGF

\exp[\lambda(e^t-1)]

26
New cards

Normal MGF

\exp \left [ \mu t + \frac{t²\sigma²}{2} \right ]

27
New cards

Gamma MGF

(1-\beta t)^{-\alpha}

28
New cards

Normal Density

p(x) = (2\pi\sigma²)^{-1/2} \exp\left [ -(2\sigma²)^{-1}(x-\mu)² \right ]

29
New cards

Binomial Density

p(x) = {n\choose x} p^x (1-p)^x

30
New cards

Binomial Expected Value

np

31
New cards

Binomial Variance

np(1-p)

32
New cards

Binomial MGF

[pe^t + (1-p)]^n

33
New cards

Chi-Square Distribution

p(x) = \frac{x^{\nu/2 - 1}e^{-x/2}}{2^{\nu/2}\Gamma(\nu/2)}

34
New cards

Chi Squared Mean

\nu

35
New cards

Chi Squared Variance

2\nu

36
New cards

Chi Squared MGF

(1-2t)^{-\nu/2}

37
New cards

P(A|B) =

P(A|B) = \frac{P(A\cap B)}{P(B)}

38
New cards

Bayes Rule: P(A|B) =

P(A|B) = P(B|A)\frac{P(A)}{P(B)}

39
New cards

Law of Total Probability

P(A_i|B) = \frac{P(B|A_i)P(A_i)}{\sum_{j=1}^{\infty} P(B|A_j)P(A_j)}

40
New cards

Independence

P(A \cap B) = P(A)P(B) AND/OR P(A|B) = P(A)

41
New cards

Random Variable

Function from a sample space into the real numbers

42
New cards

Cumulative Distribution Function

F(x) = P(X \leq x)

43
New cards

Transformations: MGF method

  1. Define U to be a function of n random variables.

  2. m_U(t) = E[e^{tu}]

  3. Match m_U(t) to mgfs of known distributions

  4. If match, then U follows known distribution by uniqueness property of MGFs

44
New cards

Transformations: Method of Distribution Functions

  1. We know Y is random var with CDF F_Y(y), and we are interested in distribution of U = h(y)

  2. Find F_U(u) using F_Y(y), remember to adjust bounds!

  3. Take derivative with respect to U to get f_U(u)

45
New cards

Transformations: Method of Transformations

  1. NEED U = h(y) to be either increasing or decreasing

  2. Find y = h^{-1}(u)

  3. Substitute y = h^{-1}(u) into f_Y(y)

  4. Multiply by \frac{dh^{-1}}{du} to find f_U(u)

  5. IF U decreasing, take absolute value

46
New cards

Expected Value

weighted average

47
New cards

Moment Generating Function

M_X(t) = E[e^{tX}] , exists if there exists a constant b such that M_X(t) is finite for |t| \leq b

48
New cards

kth moment from MGF

\mu’_k = \frac{d^k m(t)}{dt^k}|_{t=0}

49
New cards

Exponential Family definition

f(x|\theta) = h(x)c(\theta)\exp\left( \sum_{i=1}^k w_i(\theta)t_i(x) \right) where h(x) \geq 0, c(\theta) \geq 0, t_1(x), \ldots, t_k(x), w_1(\theta),\ldots, w_k(\theta) real valued.

50
New cards

Location-Scale Family general idea

specify a single pdf as the “standard”, then generate any other pdf by transforming the standard in a prescribed way

51
New cards

Location-Scale Family definition

Let f(x) be any pdf. Then, the family of pdfs \frac{1}{\sigma}f\left(\frac{x-\mu}{\sigma}\right) is the location-scale family with standard pdf f(x), location parameter \mu, and scale parameter \sigma

52
New cards

Chebychev’s Inequality

Let X be a random variable and let g(x) be a nonnegative function. Then, for any r>0, P(g(X) \geq r) \leq E(g(X))/r

53
New cards

If X \sim Pois(\lambda) then P(X = x+1) =

P(X = x+1) = \frac{\lambda}{x+1}P(X=x)

54
New cards

Integration by Parts

\int udv = uv - \int vdu

55
New cards

Marginal Distribution of X

f(x) = \int_y f(x,y)dy

56
New cards

MGF of X+Y when X,Y independent

m_{X+Y}(t) = m_X(t)m_Y(t)

57
New cards

Covariance definition

Cov(X,Y) = E[(X-\mu_X)(Y-\mu_Y)]

58
New cards

Covariance shortcut

Cov(X,Y) = E(XY) - E(X)E(Y)

59
New cards

Cov(X+Y, Z) =

Cov(X+Y, Z) = Cov(X,Z) + Cov(Y,Z)

60
New cards

Correlation definition

Cor(X,Y) = \frac{Cov(X,Y)}{\sigma_X\sigma_Y}

61
New cards

Var(\sum a_ix_i) =

Var(\sum a_ix_i) = \sum a_i² Var(x_i) + 2\sum\sum_{i < j} a_ia_jCov(X_i, X_j)

62
New cards

X_1, X_2, \ldots converges in probability to a random variable X if

for every \epsilon > 0, \lim_{n\rightarrow\infty} P(|X_n - X| < \epsilon) = 1

63
New cards

Weak Law of Large Numbers

Let X_1, X_2, \ldots be iid random variables with E(X_i) = \mu, Var(X_i) = \sigma² < \infty . Then, for every \epsilon > 0, \lim_{n \rightarrow\infty}P(|\bar{X} - \mu| < \epsilon) = 1

64
New cards

Consistent estimator

estimator is consistent for a quantity if it converges in probability to the quantity

65
New cards

If X_1, X_2, \ldots converges in probability to RV X and h(x) is a continuous function, then

If X_1, X_2, \ldots converges in probability to RV X and h(x) is a continuous function, then h(X_1), h(X_2), \ldots converges in probability to RV h(X)

66
New cards

Convergence almost surely

A sequence of random variables X_1, X_2, \ldots such that for every \epsilon > 0, P(\lim_{n\rightarrow\infty}|X_n - X| < \epsilon) = 1

67
New cards

Strong Law of Large Numbers

Let X_1, X_2, \ldots be iid random variables with E(X_i) = \mu and Var(X_i) = \sigma²<\infty. Then for every \epsilon > 0, P(\lim_{n \rightarrow\infty}|\bar{X} - \mu| < \epsilon) = 1

68
New cards

Convergence in Distribution

A sequence of random variables X_1, X_2, \ldots, such that \lim_{n\rightarrow\infty} F_{X_n}(x) = F_X(x) at all points x where F_X(x) is continuous

69
New cards

Relationship between types of convergence

a.s. \rightarrow prob \rightarrow dist

70
New cards

Central Limit Theorem

Let X_1, X_2, \ldots be a sequence of iid random variables with E[X_i] = \mu and 0< Var(X_i) = \sigma² < \infty. Then, \sqrt{n}(\bar{X} - \mu)/\sigma converges in distribution to the N(0,1) distribution

71
New cards

Slutsky’s Theorem

If X_n \rightarrow X in distribution and Y_n \rightarrow a in probability where a is a constant, then

  1. Y_nX_n \rightarrow aX in distribution

  2. X_n + Y_n \rightarrow X + a in distribution

72
New cards

Delta Method general use

When we are interested in the distribution of some function of a random variable

73
New cards

rth order Taylor Expansion about a

T_r(x) = \sum_{i=1}^r \frac{g^{(i)}(a)}{i!}(x-a)^i

74
New cards

Delta Method

Let Y_n be a sequence of ranomd variables that satisfies \sqrt{n}(Y_n - \theta) \rightarrow N(0,\sigma²) in distribution. For a given function g and a specific value of \theta, suppose that g’(\theta) exists and is nonzero. Then,

\sqrt{n}|g(Y_n) - g(\theta)| \rightarrow N(0,\sigma²[g’(\theta)]²) in distribution