Probability

Uncertainty:
observed variables (evidence)
unobserved variables
model
agent knows something about how the known variables relate to the unknown variables
Random variables:
some aspect of the world about which we (may) have uncertainty
R = is it raining?
T = is it hot or cold?
D = how long does it take to drive to work?
L = where is the ghost?
We denote with capital letters
Have domains
R = {true, false}
T = {hot, cold}
D = {0, infinity}
L = {(0,0), (2, 3)}
Probability Distributions
associate a probability with each value
unobserved random variables have distributions
ex. P(T): T P Where P(Hot) = 0.5
Hot 0.5
Cold 0.5
Joint Distributions
over a set of random variables: X1, X2, … Xn
specifies a real number for each assignment (or outcome)
P(X1 = x1, X2 = x2 … Xn = xn OR P(x1, x2 …, xn)
Size of distribution if n variables with domain size d?
dn
Probabilistic Models
a joint distribution over a set of random variables
(Random) variables with domains
assignments are called outcomes
say whether assignments are likely
normalized: sum to 1
Events:
a set of E outcomes
P(E) = ∑ P(x1, x2 …, xn)
from a joint distribution, we can calculate the probability of any event
Marginal Distributions
sub-tables which eliminate variables
Marginalization (summing out)
combining collapsed rows by adding
Conditional Probabilities
simple relation between joint and conditional probabilities
P(a | b) = P(a, b) / P(b)
Conditional Distributions
probability distributions over some variables given fixed values of others
ex. P(W | T = hot)
capital letter is an array of all the values
Normalization Trick
Step 1: select the joint probabilities matching the evidence
Step 2: normalize the selection (make it sum to 1) by dividing
Probabilistic Inference:
compute a desired probability from other known probabilities
generally compute conditional probabilities
probabilities change with new evidence

The Product Rule
P(y)P(x | y) = P(x, y)
The Chain Rule
P(x1, x2, x3) = P(x1)P(x2 | x1)P(x3 | x1 ⋅ x2)
Bayes Rule
P(x, y) = P(x | y)P(y) = P(y | x)P(x)
P(x | y) = [ P(y | x) / P(y) ] * P(x)