Title: Bayesian Networks Lecture 1: Basics and Knowledge-Based Construction
1Bayesian Networks Lecture 1Basics and
Knowledge-Based Construction
Lecture 1 is based on David Heckermans Tutorial
slides. (Microsoft Research)
Requirements 50 home works 50 Exam or a
project
2What I hope you will get out ofthis course...
- What are Bayesian networks?
- Why do we use them?
- How do we build them by hand?
- How do we build them from data?
- What are some applications?
- What is their relationship to other models?
- What are the properties of conditional
independence that make these models appropriate? - Usage in genetic linkage analysis
3Applications of hand-built Bayes Nets
- Answer Wizard 95, Office Assistant 97,2000
- Troubleshooters in Windows 98
- Lymph node pathology
- Trauma care
- NASA mission control
Some Applications of learned Bayes Nets
- Clustering users on the web (MSNBC)
- Classifying Text (spam filtering)
4Some factors that support intelligence
- Knowledge representation
- Reasoning
- Learning / adapting
5Artificial Intelligence
6Artificial Intelligence is better than none !
7Artificial Intelligence is better than ours !
8Outline for today
- Basics
- Knowledge-based construction
- Probabilistic inference
- Applications of hand-built BNs at Microsoft
9Bayesian Networks History
- 1920s Wright -- analysis of crop failure
- 1950s I.J. Good -- causality
- Early 1980s Howard and Matheson, Pearl
- Other names
- directed acyclic graphical (DAG) models
- belief networks
- causal networks
- probabilistic networks
- influence diagrams
- knowledge maps
10Bayesian Network
p(f)
p(b)
p(gf,b)
p(tb)
p(sf,t)
Directed Acyclic Graph, annotated with prob
distributions
11BN structure Definition
- Missing arcs encode independencies such that
12Independencies in a Bayes net
Example
Many other independencies are entailed by ()
can be read from the graph using d-separation
(Pearl)
13Explaining Away and Induced Dependencies
"explaining away" "induced dependencies"
14Local distributions
Table p(SyTn,Fe) 0.0 p(SyTn,Fn)
0.0 p(SyTy,Fe) 0.0 p(SyTy,Fn) 0.99
15Local distributions
Tree
16Lots of possibilities for a local distribution...
- y discrete node any probabilistic classifier
- Decision tree
- Neural net
- y continuous node any probabilistic regression
model - Linear regression with Gaussian noise
- Neural net
17Naïve Bayes Classifier
discrete
18Hidden Markov Model
discrete, hidden
H1
H2
H3
H4
H5
...
...
X1
X2
X3
X4
X5
observations
19Feed-Forward Neural Network
X1
X1
X1
inputs
hidden layer
sigmoid
Y1
Y2
Y3
outputs (binary)
sigmoid
20Outline
- Basics
- Knowledge-based construction
- Probabilistic inference
- Decision making
- Applications of hand-built BNs at Microsoft
21Building a Bayes net by hand(ok, now we're
starting to be Bayesian)
- Define variables
- Assess the structure
- Assess the local probability distributions
22What is a variable?
- Collectively exhaustive, mutually exclusive values
Error Occured
No Error
23Clarity Test Is the variable knowable in
principle
- Is it raining? Where, when, how many inches?
- Is it hot? T ? 100F , T lt 100F
- Is users personality dominant or submissive?
numerical result of standardized personality
test
24Assessing structure(one approach)
- Choose an ordering for the variables
- For each variable, identify parents Pai such that
25Example
Fuel
26Example
Fuel
p(f)
27Example
Fuel
p(bf)p(b)
p(f)
28Example
Fuel
p(bf)p(b)
p(f)
p(tb,f)p(tb)
29Example
Fuel
p(bf)p(b)
p(f)
p(tb,f)p(tb)
p(gf,b,t)p(gf,b)
30Example
p(bf)p(b)
p(f)
p(tb,f)p(tb)
p(gf,b,t)p(gf,b)
p(sf,b,t,g)p(sf,t)
p(f,b,t,g,s) p(f) p(b) p(tb) p(gf,b) p(sf,t)
31Why is this the wrong way?Variable order can be
critical
32A better wayUse causal knowledge
33Conditional Independence Simplifies Probabilistic
Inference
34Online Troubleshooters
35Define Problem
36Gather Information
37Get Recommendations
38Portion of BN for print troubleshooting
(see Breese Heckerman, 1996)
39Office Assistant 97
40Lumière Project
(see Horvitz, Breese, Heckerman, Hovel Rommelse
1998)
Users Goals
Users Needs
User Activity
41Studies with Human Subjects
- Wizard of OZ experiments at MS Usability Labs
User Actions
Typed Advice
Expert Advisor
Inexperienced user
42Activities with Relevance to Users Needs
Several classes of evidence
- Search e.g., menu surfing
- Introspection e.g., sudden pause, slowing of
command stream - Focus of attention e.g, selected objects
- Undesired effects e.g., command/undo, dialogue
opened and cancelled - Inefficient command sequences
- Goal-specific sequences of actions
43Summary so far
- Bayes nets are useful because...
- They encode independence explicitly
- more parsimonious models
- efficient inference
- They encode independence graphically
- Easier explanation
- Easier encoding
- They sometimes correspond to causal models
- Easier explanation
- Easier encoding
- Modularity leads to easier maintenance
44Teenage Bayes
MICRONEWS 97 Microsoft Researchers Exchange
Brainpower with Eighth-grader Teenager Designs
Award-Winning Science Project .. For her science
project, which she called "Dr. Sigmund
Microchip," Tovar wanted to create a computer
program to diagnose the probability of certain
personality types. With only answers from a few
questions, the program was able to accurately
diagnose the correct personality type 90 percent
of the time.
45Artificial Intelligence is a promising
fieldalways was, always will be.