uncertainty-probabilty
uncertainty-probabilty
uncertainty-probabilty
Outline
• Uncertainty
• Probability
• Syntax and Semantics
• Inference
• Independence and Bayes' Rule
Uncertainty
Most intelligent systems have some degree of uncertainty.
Uncertainty in systems arises primarily because of problems in data.
Such as missing data or unavailable data, data is present but
unreliable or ambiguous due to errors in measurement, presence of
multiple conflicting measurements etc.
Always it is not possible to represent data in a precise and consistent
manner; data is generally based on defaults, which may have
exceptions leading to errors in intelligent systems.
Uncertainty may also be caused by the represented knowledge since it
might represent only the best guesses of the expert based on
observations or statistical conclusion which may not be appropriate
in all situation.
Methods for handling uncertainty : probability theory, fuzzy logic, model
and temporal logic.
Uncertainty
Let action At = leave for airport t minutes before flight.
Will At get me there on time?
Problems:
1. partial observability (road state, other drivers' plans, etc.)
2. noisy sensors (traffic reports)
3. uncertainty in action outcomes (flat tire, etc.)
4. immense complexity of modeling and predicting traffic
“A25 will get me there on time if there's no accident on the bridge and it
doesn't rain and my tires remain intact etc etc.”
(A1440 might reasonably be said to get me there on time but I'd have to
stay overnight in the airport …)
Methods for handling uncertainty
• Default or nonmonotonic logic:
– Assume my car does not have a flat tire
– Assume A25 works unless contradicted by evidence
• Issues: What assumptions are reasonable? How to handle
contradiction?
Subjective probability:
• Probabilities relate propositions to agent's own state of
knowledge
e.g., P(A25 | no reported accidents) = 0.06
These are not assertions about the world
Then the required summation of joint entries is done by summing out the
hidden variables:
P(Y | E = e) = αP(Y,E = e) = αΣhP(Y,E= e, H = h)
• The terms in the summation are joint entries because Y, E and H together
exhaust the set of random variables
• Obvious problems:
1. Worst-case time complexity O(d n) where d is the largest arity
2. Space complexity O(dn) to store the joint distribution
3. How to find the numbers for O(dn) entries?
Independence
• A and B are independent iff
P(A|B) = P(A) or P(B|A) = P(B) or P(A, B) = P(A) P(B)
• Equivalent statements:
P(Toothache | Catch, Cavity) = P(Toothache | Cavity)
P(Toothache, Catch | Cavity) = P(Toothache | Cavity) P(Catch | Cavity)
Conditional independence
contd.
• Write out full joint distribution using chain rule:
P(Toothache, Catch, Cavity)
= P(Toothache | Catch, Cavity) P(Catch, Cavity)
= P(Toothache | Catch, Cavity) P(Catch | Cavity) P(Cavity)
= P(Toothache | Cavity) P(Catch | Cavity) P(Cavity)
I.e., 2 + 2 + 1 = 5 independent numbers
• In most cases, the use of conditional independence
reduces the size of the representation of the joint
distribution from exponential in n to linear in n.
• or in distribution form
P(Y|X) = P(X|Y) P(Y) / P(X) = αP(X|Y) P(Y)