4: Dimensional Analysis by Statistical Analysis

0.0(0)
studied byStudied by 0 people
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
Card Sorting

1/26

encourage image

There's no tags or description

Looks like no tags are added yet.

Study Analytics
Name
Mastery
Learn
Test
Matching
Spaced

No study sessions yet.

27 Terms

1
New cards

Probability Function

Random variable X, f(xi) = Pr(X = xi)

2
New cards

Dense Sampling Set

A sampling space that’s uncountable and continuous.

3
New cards

Probability Density Function

For every closed interval xi = [a, b], where for all x, f(x) >= 0 and the integral of f(x) between negative infinity and infinity is 1.

  • S is a dense sampling space

  • Powerset(S) = {ci} and the powerset of all subsets of S

  • X: S → T is a dense random variable defined over S where T = {xi} and X(ci) = xi

<p>For every closed interval x<sub>i</sub> = [a, b], where for all x, f(x) &gt;= 0 and the integral of f(x) between negative infinity and infinity is 1.</p><ul><li><p>S is a dense sampling space</p></li><li><p><span>Powerset(S) = {c<sub>i</sub>} and the powerset of all subsets of S</span></p></li><li><p><span>X: S → T is a dense random variable defined over S where T = {x<sub>i</sub>} and X(c<sub>i</sub>) = x<sub>i</sub></span></p></li></ul><p></p>
4
New cards

Probability Density Function Properties

  • For all x, Pr(X = x) = 0

  • P(X >= a) = integral between infinity and a of f(x)

  • P(X <= a) = integral between a and negative infinity of f(x)

5
New cards

Cumulative Probability Distribution Function

F(x) = P(X <= x)

  • X is a random variable defined over a sampling space

6
New cards

Cumulative Probability Distribution Properties

  • Always increasing but not always monotonic - x1 < x2 → F(x1) <= F(x2)

  • lim(x → -inf.) F(x) = 0 and lim(x → inf.) F(x) = 1

  • Pr(X > x) = 1 - F(x)

  • x1 < x2 → Pr(x1 < X <= x2) = F(x2) - F(x1)

    • Note - F(x) is only continuous on the right-hand side, but not always on the left

7
New cards

Joint Distribution

A collection of probabilities over a series of random variables’ sampling spaces: Pr(X1 = cx1, j, …, Xn = cxn, k)

  • X are the random variables

  • n is the number of random variables

  • Powerset(Xi) = {cxi}

  • j and k indicate elements in x

Denoted as f(x, y)

8
New cards

Bivariate Distribution

Joint distribution where n = 2 (i.e. there are two random variables)

9
New cards

Contingency Tables

Frequency matrices that express joint frequences for 2 or more categorial variables.

<p>Frequency matrices that express joint frequences for 2 or more categorial variables.</p>
10
New cards

Mutual exclusion (Disjoint)

  • Where two sets never contain common outcomes (A ∩ B = {}).

  • For more than two sets, for each pair of sets denoted by i, j where i ≠ j, Ai ∩ Aj = {}

11
New cards

Probability of a union of events

knowt flashcard image
12
New cards

Probability of a union of disjoint events

Just a sum of all individual probabilities.

13
New cards

Probability of a union of two sets

P(A ∩ B) = P(A) + P(B) - P(A U B)

14
New cards

Independence of two sets

P(A ∩ B) = P(A) * P(B)

  • The occurrence of one event does not give us information about another event.

15
New cards

Statistically Independent Random Variables

The joint probability distribution function is factorisable in the form P(X, Y) = P(X) * P(Y)

16
New cards

Information contained in the occurrence of an event

I(ei) = -logb(P(ei)) = logb(1/P(ei))

17
New cards

Units of information

Depends on the base used:

  • Bits if 2

  • Hartleys if 10

  • Nats if e

18
New cards

Entropy

The average amount of information you get from the source - a measure of uncertainty.

  • Affected by statistical independence

  • The more common the event is, the less information it carries

19
New cards

Information (r.e. Entropy)

Resolves uncertainty, tells us more about a random variable.

20
New cards

Entropy Formula

H(X) = For all i, P(xi) * I(xi) = -(for all i, P(xi) * logbP(xi))

21
New cards

Joint Entropy Formula

H(X, Y) = -(for all xi in X (for all yj in Y (P(xi, yj) * logbP(xi, yj))))

22
New cards

Independent Component Analysis

Used to decompose a signal into statistically independent sources (bases).

  • In the formula Y = XB, ICA solves for both X and B at the same time.

  • Mathematically solving X ~= X^ = WY, W ~= B-1

  • Only produces approximations

23
New cards

Independent Component Analysis Formula (Long)

  • Observe J linear mixtures y1, … yj comprised of I = J sources x1, …, xi

  • For all j in J, B(1, j)x1 + B(2, j)x2 + … + B(i, j)xi

24
New cards

Independent Component Analysis Formula

Y = XB = for all i in I (sources), B(i, j)xi

25
New cards

Independent Component Analysis Limitations

  • Always assumes the components are mutually independent, mean-centred and have non-Gaussian distributions

  • Cannot identify the number of source signals or the proper scaling of them

  • Ignores sampling order (time dependency) and works over variables rather than signals

26
New cards

Solving for an estimation - assumption (ICA)

That the sources are mutually exclusive.

27
New cards

Solving for an estimation (ICA)

Maximising the joint entropy of X^, i.e. g(X^) = g(WY)

  • g(X^) is the cumulative density function of X^

  • Independence of sources is obtained by adjusting the mixing of W ~= B-1

  • This also maximises the mutual entropy, H(g(X^))

  • Then use gradient descent to take a small step in the direction of the gradient of H(g(X^))

  • Wnew = Wold + h * gradient(H(g(X^)))

    • h is the learning rate