Probability and Statistics Refresher (Week1)

0.0(0)
studied byStudied by 0 people
GameKnowt Play
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/39

flashcard set

Earn XP

Description and Tags

Quants

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

40 Terms

1
New cards

Random Variables ?

A variable which takes on a numerical value as a result of a random process

Discrete R.V: The number of outcomes of the random variable are distinct and countable

Continuos R.V: The random variable can take on infinitely many different outcomes

2
New cards

How are Random variables treated in Quant Ecos?

Random variables are treated as continous even though they may technically be discrete

3
New cards

Probability density Functions (PDF’s)

3 properties for a valid PDF:

  • f(x) > 0

  • The integral of f(x)dx = 1

<p>3 properties for a valid PDF:</p><ul><li><p>f(x) &gt; 0</p></li><li><p>The integral of f(x)dx = 1</p></li></ul><p></p>
4
New cards

Bernoulli Random Variable

Special type of discrete random variable with 2 possible outcomes. e.g coin toss

It is descibed by the probability of success (observing 1)

<p>Special type of discrete random variable with 2 possible outcomes. e.g coin toss </p><p>It is descibed by the probability of success (observing 1)</p>
5
New cards

Discrete Random Variable with more than 2 outcomes

If you have more than 2 outcomes, just list the probability that the random variable takes on each of the values

6
New cards

Continuos Random Variables

With CT RV’s, we have infinitely many x-values that our variable can take on. SO the probability of taking on an exact value is 0. Hence, we look at the area between 2 points

<p>With CT RV’s, we have infinitely many x-values that our variable can take on. SO the probability of taking on an exact value is 0. Hence, we look at the area between 2 points </p>
7
New cards

Cumulative Distribution Functions (CDF)

Defined the same for both discrete and continuous rv’s.

F(X) = Pr(X<x) : Which is the probability of observing an outcome as large as x

8
New cards

CDF Properties

knowt flashcard image
9
New cards

Using CDF’s to calculate probabilities

2 properties:

  • Pr(x>b) = 1 - F(b)

  • Pr (a<X<b) = F(b) - F(a)

10
New cards

Joint Probabilities

When we have 2 random variables, X & Y, we have a joint distribution

The distribution describes the probability of 2 events happening together.

Joint PDF: f(X,Y) = Pr(X=x,Y=y)

11
New cards

Joint PDF properties

For a joint pdf to be valid:

  • joint probabiity values should lie between 0 and 1

  • joint probabilites need to sum to 1

<p>For a joint pdf to be valid:</p><ul><li><p>joint probabiity values should lie between 0 and 1</p></li><li><p>joint probabilites need to sum to 1</p></li></ul><p></p>
12
New cards

Marginal Distributions

A marginal distribution is the probability distribution of a single variable in a multivariable (usually joint) distribution. It is found by summing (for discrete variables) or integrating (for continuous variables) over the other variable(s), effectively "marginalizing out" the others.

<p>A marginal distribution is the probability distribution of a single variable in a multivariable (usually joint) distribution. It is found by summing (for discrete variables) or integrating (for continuous variables) over the other variable(s), effectively "marginalizing out" the others.</p>
13
New cards

Statistical Independence

Two events are statistically independent if:

  • Pr(A and B) = Pr (A) x Pr(B)

In the case of 2 random variables, X and Y are statistically independent iff their joint probability is the product of their respective marginal distributions

14
New cards

Proving Statistical indepenence

  • To prove statistical independence, we have to show the property holds for all cells.

  • To disprove, we need to only show one that fails - then they are statistically dependent

15
New cards

Conditional Distributions

When you want to determine how one of the variables changes if we fix the other one at a specific level.

  • Note: If you hold an event fixed, all the distributions containing that event should sum to 1

<p>When you want to determine how one of the variables changes if we fix the other one at a specific level. </p><ul><li><p>Note: If you hold an event fixed, all the distributions containing that event should sum to 1</p></li></ul><p></p>
16
New cards

Conditional distributions and statistical Independence

A conditional distribution will tell you the probability distribution of one random variable, whilst holding the other fixed.

If X and Y are statistically independent, then the cond.distr of x given a value of Y should be the same no matter what we choose to fix Y on.

17
New cards

Expected Value operator E(x)

A measure of central tendency. It gives us a measure of the value we expect a random variable to take on. It is also the 1st moment of the distribution.

Note: The expected value of a random variable does not need to be an exact value that the random variable can take on.

<p>A measure of central tendency. It gives us a measure of the value we expect a random variable to take on. It is also the 1st moment of the distribution.</p><p>Note: The expected value of a random variable does not need to be an exact value that the random variable can take on.</p>
18
New cards

Properties of Expected Values E(x)

  • If c is a constant, then: E(c)= c

  • If a & b are constants, then:

    E(aX+b) = aE(x) + b

19
New cards

Functions of random variables

Note: If X is a random variable, then X² is also a random variable etc.

  • Applying any function to a random variable only changes the outcome with which the random variable can take on, and not the probability distribution of the random variable’s value.

<p>Note: If X is a random variable, then X² is also a random variable etc.</p><ul><li><p>Applying any function to a random variable only changes the outcome with which the random variable can take on, and not the probability distribution of the random variable’s value. </p></li></ul><p></p>
20
New cards

Measures of variability: Variance

Variance - A measure of spread; how far a random variable deviates from its mean. It is also the 2nd moment of the distribution

Var(x) = [E(X-E(X))²] = E(X²) - (E(X))²

21
New cards

Properties of Variance

  • Variance of a constant is zero as a constant does not vary at all

  • Var(aX + b) = a²var(X)

22
New cards

Standard deviation and its properties

The positive square root of the variance of a random variable. Written as sd(x).

Properties:

  • for constant c, sd(c) = 0

  • for constants a,b; sd(aX +b) = |a|sd(x)

23
New cards

Standardizing a Random Variable

To standardize a variable to obtain a random variable with a distribution with a mean of zero and variance of 1, we subtract by its mean and divide by the standard deviation.

<p>To standardize a variable to obtain a random variable with a distribution with a mean of zero and variance of 1, we subtract by its mean and divide by the standard deviation. </p>
24
New cards

Skewness and Kurtosis

Skewness is the 3rd moment of a distribution. Depicts how skew a random variable is.

Kurtosis is the 4th moment of a distribution. Depcits the thickness of the tails of the distribution.

25
New cards

Relationship between covariance & correlation

Linked in the sense that the correlation between 2 variables is simply the covariance adjusted for the standard deviations of those variables.

Both concepts refer to the notion of how 2 variables move together

26
New cards

What is covariance

The covariance between 2 random variables measures the amount of linear dependence between 2 random variables. Written as, Cov(X, Y)

Note:

  • +’ve Cov: X and Y move in the same direction

  • -’ve Cov: X and Y move in opposite directions

Cov(X,Y) = E(XY) - E(X)E(Y)

Cov(X,X) = Var(X)

27
New cards

Properties of Covariance

  • If X & Y are independent, then Cov(X,Y) = 0, however the reverse is not always true

  • Cov(a1X+b1, a2Y+b2) = a1a2Cov(X,Y)

  • |Cov(X,Y)| < sd(X)sd(Y)

28
New cards

The Correlation Coefficient

Corr(X,Y) is the covariance betwen X and Y standardized by their standard deviations.

<p>Corr(X,Y) is the covariance betwen X and Y standardized by their standard deviations. </p>
29
New cards

Properties of Correlation

  • -1 < Corr(X,Y) < 1

  • if a1,a2 > 0, then Corr(a1X+b1,a2Y+b2) = Corr(X,Y)

  • if a1,a2 <0, then Corr(a1X+b1,a2X+b2) = -Corr(X,Y)

    Reflects that the Correaltion coefficient is scale independent

30
New cards

Variances of sums of Random Variables

What if you want to calculate the variance of a newly created random variable due to a sum of 2 random variables?

Var(aX + bY) = a²Var(X) + b²Var(Y) + 2abCov(X,Y)

  • If x and y are uncorrelated, then the covariance term drops

31
New cards

Conditional Expectation

The conditional expectation operator is one that calculates the expected value of one random variable, while holding the other fixed at some particular value.

<p>The conditional expectation operator is one that calculates the expected value of one random variable, while holding the other fixed at some particular value.</p>
32
New cards

Properties of condtional expectations

  • E[c(x) | X] = c(X); functions of X are treated as constants when we control for X

  • If X and Y are independent, then E(Y|X) = E(Y), then Cov(X,Y) = 0

  • For functions a(X) and b(X), E[a(X)Y + b(X)|X] = a(X)E(Y|X) + b(X)

  • Ex[E(Y|X)] = E(Y)

33
New cards

Conditional Variance

Var(Y|X=x) = E(Y²|x) - [E(Y|x)]²

  • If x and y are independent, then Var(Y|X) = Var(Y)

34
New cards

The normal distribution

Characteristics of a normal distribution:

  • skewness = 0 and kurtosis = 3

  • STD normal will have a mean of 0 and varinace of 1

  • X~Normal(mew, sigma²)

35
New cards

The standard normal table

Calculating probabilities using tables:

  • Have to standardize your random variable first

  • Area under a std normal curve is 1

  • The table gives the area under the curve to the left of your particular number

36
New cards

Estimates and Estimators

An estimator of a parameter theta is a rule that assigns each possible sample a value of theta.

Once you apply the rule to a particular data, you obtain an estimate.

<p>An estimator of a parameter theta is a rule that assigns each possible sample a value of theta. </p><p>Once you apply the rule to a particular data, you obtain an estimate. </p>
37
New cards

Unbiasedness

How well does an estimator estimate the true parameter value?

We can check by determining whether an estimator is unbiased or not:

  • An estimator, W, of a parameter theta is unbiased if E(W) = theta

  • If it ‘s biased, then E(W) = theta + Bias(W)

    Note: Bias(W) = E(W) - theta

38
New cards

Sampling Variances

Unbiasedness alone is not good enough to conclude whether an estimator is good or not;

Thus another metric of how precise an estimator predicts the populaton parameter is the sample variance of the estimator.

  • What is the spread of values the estimator gives us.

39
New cards

Efficiency of an Estimator

We want to determine which of our estimators are more useful than others. Naturally we want our estimators to be unbiased and precise as possible. Hence, ideally the lower the sampling variance, the better.

Efficiency:

  • If Var(W1) < Var(W2), the W1 estimator is more efficient

40
New cards

Consistency

What happens to your sampling distribution of your estimator as you increase the size of the sample you draw from the population.

An estimator is consistent if:

  • It is unbiased

  • The sampling variance goes to zero as the sample size (n) increases.