0% found this document useful (0 votes)
68 views

Good BayesianNetworksPrimer

Bayesian networks are graphical models that represent probabilistic relationships among variables. A Bayesian network consists of a directed acyclic graph whose nodes represent variables and edges represent conditional dependencies. Each node has a conditional probability table that quantifies the probabilistic effects of its parent nodes. Bayesian networks provide a compact representation of joint probability distributions and allow efficient inference and learning of causal relationships from data.

Uploaded by

pradnya kinge
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
68 views

Good BayesianNetworksPrimer

Bayesian networks are graphical models that represent probabilistic relationships among variables. A Bayesian network consists of a directed acyclic graph whose nodes represent variables and edges represent conditional dependencies. Each node has a conditional probability table that quantifies the probabilistic effects of its parent nodes. Bayesian networks provide a compact representation of joint probability distributions and allow efficient inference and learning of causal relationships from data.

Uploaded by

pradnya kinge
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 23

Bayesian Networks

ABRIEFINTRODUCTIO
N

ADNAN MASOOD
SCIS.NOVA.EDU/~ADNAN
ADNAN@NOVA.EDU

DOCTORALCANDIDATE
NOVASOUTHEASTERN UNIVERSITY
What is a Bayesian Network?

 A Bayesian network (BN) is a graphical model for


depicting probabilistic relationships among a set
of variables.
 BN Encodes the conditional independence relationships between the
variables in the graph structure.
 Provides a compact representation of the joint probability
distribution over the variables
 A problem domain is modeled by a list of variables X1, …, Xn
 Knowledge about the problem domain is represented by a joint
probability P(X1, …, Xn)
 Directed links represent causal direct influences
 Each node has a conditional probability table quantifying the effects
from the parents.
 No directed cycles
Bayesian Network constitutes of..

 Directed Acyclic Graph (DAG)


 Set of conditional probability tables for each node in
the graph

C D
So BN = (DAG, CPD)

 DAG: directed acyclic graph (BN’s structure)


 Nodes: random variables (typically binary or discrete,
but methods also exist to handle continuous variables)
 Arcs: indicate probabilistic dependencies between
nodes (lack of link signifies conditional independence)

 CPD: conditional probability distribution (BN’s


parameters)
 Conditional probabilities at each node, usually stored
as a table (conditional probability table, or CPT)
So, what is a DAG?

Follow the general graph


directed acyclic graphs use principles such as a node A is a
only unidirectional arrows to parent of another node B, if
show the direction of there is an arrow from node A
causation to node B.

B Informally, an arrow from


node X to node Y means X has
a direct influence on Y

C D
Each node in graph represents
a random variable
Where do all these numbers come from?
There is a set of tables for each node in the network.
Each node Xi has a conditional probability distribution
P(Xi | Parents(Xi)) that quantifies the effect of the parents
on the node
The parameters are the probabilities in these conditional
A probability tables (CPTs)

C D
The infamous Burglary-Alarm Example

Burglary
P(B) Earthquake
P(E)
0.001 0.002
B E P(A)
T T 0.95
Alarm T F 0.94
F T 0.29
F F 0.001
A P(J)
John Calls T 0.90 Mary Calls A P(M)
F 0.05 T 0.70
F 0.01
Cont..calculations on the belief network

Using the network in the example, suppose you want


to calculate:
P(A = true, B = true, C = true, D = true)
This is from the
= P(A = true) * P(B = true | A = true) * graph structure
P(C = true | B = true) P( D = true | B = true)
= (0.4)*(0.3)*(0.1)*(0.95)

These numbers are from the


conditional probability tables
So let’s see how you can calculate P(John called)
if there was a burglary?

Inference from effect to cause; Given a burglary,


what is P(J|B)?
P(J | B)  ?
P( A | B)  P(B)P(E)(0.94)  P(B)P(E)(0.95)
P( A | B)  1(0.998)(0.94)  1(0.002)(0.95)
P( A | B)  0.94
P(J | B)  P( A)(0.9)  P(A)(0.05)
P(J | B)  (0.94)(0.9)  (0.06)(0.05)
 0.85
Can also calculate P (M|B) = 0.67
Why Bayesian Networks?

Bayesian Probability represents the degree of belief


in that event while Classical Probability (or
frequents approach) deals with true or physical
probability of an event
• Bayesian Network
• Handling of Incomplete Data Sets
• Learning about Causal Networks
• Facilitating the combination of domain knowledge and data
• Efficient and principled approach for avoiding the over fitting
of data
What are Belief Computations?

 Belief Revision
 Model explanatory/diagnostic tasks
 Given evidence, what is the most likely hypothesis to explain the
evidence?
 Also called abductive reasoning
 Example: Given some evidence variables, find the state of all other
variables that maximize the probability. E.g.: We know John Calls,
but not Mary. What is the most likely state? Only consider
assignments where J=T and M=F, and maximize.
 Belief Updating
 Queries
 Given evidence, what is the probability of some other random
variable occurring?
What is conditional independence?

The Markov condition says that given its parents (P1, P2), a
node (X) is conditionally independent of its non-descendants
(ND1, ND2)

P1 P2

ND1 X ND2

C1 C2
What is D-Separation?

 A variable a is d-separated from b by a set of variables


E if there does not exist a d-connecting path between a
and b such that
 None of its linear or diverging nodes is in E
 For each of the converging nodes, either it or one of its
descendants is in E.
 Intuition:
 The influence between a and b must propagate through a d-
connecting path
 If a and b are d-separated by E, then they are
conditionally independent of each other given E:
P(a, b | E) = P(a | E) x P(b | E)
Construction of a Belief Network

Procedure for constructing BN:


 Choose a set of variables describing the application
domain
 Choose an ordering of variables
 Start with empty network and add variables to the
network one by one according to the ordering
 To add i-th variable Xi:
 Determine pa(Xi) of variables already in the network (X1, …,
Xi – 1)
such that
P(Xi | X1, …, Xi – 1) = P(Xi | pa(Xi))
(domain knowledge is needed there)
 Draw an arc from each variable in pa(Xi) to Xi
What is Inference in BN?

 Using a Bayesian network to compute probabilities is


called inference
 In general, inference involves queries of the form:
P( X | E )
where X is the query variable and E is the evidence
variable.
Representing causality in Bayesian Networks

 A causal Bayesian network, or simply causal


networks, is a Bayesian network whose arcs are
interpreted as indicating cause-effect relationships
 Build a causal network:
 Choose a set of variables that describes the domain
 Draw an arc to a variable from each of its direct causes
(Domain knowledge required)
Visit Africa Lung Cancer
Smoking

Tuberculosis

Tuberculosis or Bronchitis
Lung Cancer

X-Ray
Dyspnea
Limitations of Bayesian Networks

• Typically require initial knowledge of many


probabilities…quality and extent of prior knowledge
play an important role
• Significant computational cost(NP hard task)
• Unanticipated probability of an event is not
taken care of.
Summary

 Bayesian methods provide sound theory and framework for


implementation of classifiers
 Bayesian networks a natural way to represent conditional independence
information. Qualitative info in links, quantitative in tables.
 NP-complete or NP-hard to compute exact values; typical to make
simplifying assumptions or approximate methods.
 Many Bayesian tools and systems exist
 Bayesian Networks: an efficient and effective representation of the joint
probability distribution of a set of random variables
 Efficient:
 Local models
 Independence (d-separation)
 Effective:
 Algorithms take advantage of structure to
 Compute posterior probabilities
 Compute most probable instantiation
 Decision making
Bayesian Network Resources

 Repository: www.cs.huji.ac.il/labs/compbio/Repository/
 Softwares:
 Infer.NET http://research.microsoft.com/en-
us/um/cambridge/projects/infernet/
 Genie: genie.sis.pitt.edu
 Hugin: www.hugin.com
 SamIam http://reasoning.cs.ucla.edu/samiam/
 JavaBayes: www.cs.cmu.edu/ javabayes/Home/
 Bayesware: www.bayesware.com

 BN info sites
 Bayesian Belief Network site (Russell Greiner)
http://webdocs.cs.ualberta.ca/~greiner/bn.html
 Summary of BN software and links to software sites (Kevin Murphy)
References and Further Reading

 Bayesian Networks without Tears by Eugene Charniak


http://www.cs.ubc.ca/~murphyk/Bayes/Charniak_91.
pdf
 Russel, S. and Norvig, P. (1995). Artificial
Intelligence, A Modern Approach. Prentice
Hall.
 Weiss, S. and Kulikowski, C. (1991). Computer Systems
That Learn. Morgan Kaufman.
 Heckerman, D. (1996). A Tutorial on Learning
with Bayesian Networks. Microsoft Technical
Report MSR-TR-95-06.
 Internet Resources on Bayesian Networks and
Machine Learning:
http://www.cs.orst.edu/~wangxi/resource.html
Modeling and Reasoning with Bayesian
Networks
Machine Learning: A Probabilistic Perspective
Bayesian Reasoning and Machine Learning

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy