Foundations of Probability and Hidden Markov Models

0.0(0)
Studied by 0 people
call kaiCall Kai
learnLearn
examPractice Test
spaced repetitionSpaced Repetition
heart puzzleMatch
flashcardsFlashcards
GameKnowt Play
Card Sorting

1/49

flashcard set

Earn XP

Description and Tags

Comprehensive practice flashcards covering probability foundations, probabilistic graphical models, HMM algorithms (Viterbi, Forward, Backward, Baum-Welch), and biological applications like GENSCAN and CpG Islands.

Last updated 6:29 AM on 4/29/26
Name
Mastery
Learn
Test
Matching
Spaced
Call with Kai

No analytics yet

Send a link to your students to track their progress

50 Terms

1
New cards

Sample Space

The set of all possible outcomes of an experiment, which can be discrete (e.g., A,C,G,T{A, C, G, T}) or continuous (e.g., (0,extN)(0, ext{N})).

2
New cards

Elementary Events

Outcomes that are atomic, mutually exclusive, and collectively exhaustive.

3
New cards

Atomic Events

Events that cannot be broken down further into smaller components.

4
New cards

Mutually Exclusive

A relationship between events where only one can occur at a time.

5
New cards

Collectively Exhaustive

A set of events whose probabilities sum to exactly 11.

6
New cards

Probability Axiom 1 (Non-negativity)

The rule stating that there are no negative probabilities, represented as P(E)extextisextalwaysextatextleast0P(E) ext{ } ext{is } ext{always } ext{at } ext{least } 0.

7
New cards

Probability Axiom 2 (Normalization)

The requirement that the sum of probabilities for all outcomes in a sample space must equal exactly 11.

8
New cards

Addition Rule (Axiom 3)

If two events are mutually exclusive, the probability of one or the other occurring is the sum of their individual probabilities (P(E1extorE2)=P(E1)+P(E2)P(E_1 ext{ or } E_2) = P(E_1) + P(E_2)).

9
New cards

Joint Probability

The probability that two different random variables take on specific values at the same time, denoted as P(X,Y)P(X, Y).

10
New cards

Marginal Probability

The probability of one variable happening regardless of the result of another, calculated by "marginalizing out" or summing across the other variable: P(X=x)=extSumoverYextofP(X=x,Y=y)P(X=x) = ext{Sum over } Y ext{ of } P(X=x, Y=y).

11
New cards

Conditional Probability

The probability that event XX happens given that event YY has already happened, denoted as P(XY)P(X|Y).

12
New cards

Independence

The property where two variables do not impact each other, mathematically defined when P(X,Y)=P(X)imesP(Y)P(X, Y) = P(X) imes P(Y) or P(XY)=P(X)P(X|Y) = P(X).

13
New cards

Bayes' Rule

A formula used for "The Great Inversion" to find inverse probability, expressed as P(S|X) = rac{P(X|S) imes P(S)}{P(X)}.

14
New cards

Prior Belief

In Bayes' Rule, the initial belief (P(S)P(S)) before seeing new data.

15
New cards

Normalizing Factor

The denominator in Bayes' Rule (P(X)P(X)) that ensures the total probability equals 11.

16
New cards

Probabilistic Graphical Models (PGMs)

A map of a joint probability distribution used to model large biological systems using assumptions.

17
New cards

Nodes

Represented as circles in a PGM, each node corresponds to a specific random variable.

18
New cards

Edges

Directed arrows in a graph indicating a conditional dependency (e.g., XightarrowYX ightarrow Y implies one must know XX for YY).

19
New cards

Roots

Nodes in a PGM with no parents, representing prior independent events.

20
New cards

Parents

In a relationship XightarrowYX ightarrow Y, XX is the parent node.

21
New cards

Children

In a relationship XightarrowYX ightarrow Y, YY is the child node.

22
New cards

Graphical Factoring Rule

The joint probability of a whole system is the product of every node given its parents: P(X1,ext,Xn)=extProductofP(XiextParents(Xi))P(X_1, ext{…}, X_n) = ext{Product of } P(X_i | ext{Parents}(X_i)).

23
New cards

Maximum Likelihood Estimation (MLE)

A process of parameter estimation using real biological data to find probabilities by counting and normalizing.

24
New cards

First Order Markov Property

The assumption that the "future" depends only on the "present" and not the past: P(Xi+1Xi,Xi1,ext,X0)=P(Xi+1Xi)P(X_{i+1} | X_i, X_{i-1}, ext{…}, X_0) = P(X_{i+1} | X_i).

25
New cards

Markov Model (Chain)

A graphical model that looks like a line of arrows used to calculate the probability of a specific sequence (e.g., DNA).

26
New cards

CpG Islands

Genomic regions near gene promoters where the methylation of Cytosine to Thymine is blocked, making "CG" sequences more common.

27
New cards

Hidden Markov Models (HMMs)

A model used when the biological function (hidden states) is not visible, but symbols (emissions) are observed.

28
New cards

Doubly Stochastic

A property of HMMs representing two levels of randomness: State Transition and Symbol Emission.

29
New cards

Initial Distribution (oldsymbol{ ext{\pi}})

The probability of starting in each specific hidden state at the beginning of a sequence where i=1i = 1.

30
New cards

Transition Matrix (oldsymbol{ ext{A}})

A KimesKK imes K matrix where each value akla_{kl} is the probability of moving from state kk to state ll, with each row summing to 11.

31
New cards

Emission Matrix (oldsymbol{ ext{E}})

A KimesBK imes B matrix where ek(b)e_k(b) is the probability that state kk emits symbol bb, with all symbols within a state summing to 11.

32
New cards

The Decoding Problem

The challenge of identifying the single most likely sequence of hidden states that produced a sequence of observations.

33
New cards

Viterbi Algorithm

A dynamic programming algorithm that solves the decoding problem in linear time using a Trellis Diagram.

34
New cards

Trellis Diagram

A graph where columns represent positions in the DNA sequence and rows represent hidden states, used by the Viterbi algorithm.

35
New cards

Viterbi Traceback

The process of following back-pointers from the final maximum score to reconstruct the path of hidden states.

36
New cards

Log-space Math

The practice of taking the logarithm of probabilities to prevent computational underflow when multiplying many tiny numbers.

37
New cards

Posterior Decoding

A local decoding method that considers all possible paths passing through a state at a specific position to find its probability.

38
New cards

Forward Algorithm

Calculates the total probability (fk(i)f_k(i)) of all paths that reach state kk at position ii by summing across previous states.

39
New cards

Backward Algorithm

Calculates probability (bk(i)b_k(i)) by working in reverse from the end of a sequence to the current position to predict the present based on the future.

40
New cards

Baum-Welch Algorithm

An expectation-maximization (EM) algorithm used to estimate HMM parameters when labeled data is not provided.

41
New cards

Expectation Step (E)

The step in Baum-Welch where the forward-backward algorithm is used to calculate the expected counts of transitions and emissions.

42
New cards

Maximization Step (M)

The step in Baum-Welch where factorial counts are used to update the Transition (A) and Emission (E) matrices.

43
New cards

GENSCAN

A generalized HMM (GHMM) developed by Burge and Karlin in 19971997 to predict gene structure in human DNA.

44
New cards

Duration Modeling

A feature of GHMMs allowing state lengths to be modeled explicitly (e.g., average exon length of 180180 bp) rather than decaying exponentially.

45
New cards

Local Maxima

A potential pitfall of the Baum-Welch algorithm where the model converges on a good solution that is not the absolute best (global) solution.

46
New cards

Ultrametricity

A property that assumes a molecular clock where all leaves in a phylogenetic tree are equidistant from the root.

47
New cards

Additivity

A property that does not assume mutations are proportionate to time, producing unrooted trees that require an outgroup.

48
New cards

Affine Gap

A 33-matrix dynamic programming approach where different penalties are applied for opening versus extending a gap.

49
New cards

Long-branch Attraction

An error where sequences with many mutations are incorrectly grouped together because they appear similar by chance.

50
New cards

NJ Algorithm

The Neighbor-Joining algorithm developed by Saitou and Nei used for building phylogenetic trees.