Soft Computing - PowerPoint PPT Presentation

About This Presentation
Title:

Soft Computing

Description:

Probabilistic is used for description of events or behavior or ... Probability theory and then probabilistic methods of AI aim ... believe/disbelieve them to ... – PowerPoint PPT presentation

Number of Views:109
Avg rating:3.0/5.0
Slides: 23
Provided by: gavr6
Category:

less

Transcript and Presenter's Notes

Title: Soft Computing


1
Soft Computing
  • Lecture 17
  • Introduction to probabilistic reasoning. Bayesian
    nets. Markov models

2
Why probabilistic methods?
  • Probabilistic is used for description of events
    or behavior or making decision when we have not
    enough knowledge about object of observation
  • Probability theory and then probabilistic methods
    of AI aim to introduce in random processes any
    knowledge about laws or rules about sources of
    random events
  • This is alternative path of description of
    uncertainty in contrast to fuzzy logics

3
The Problem
  • We normally deal with assertions and their causal
    connections
  • John has fever
  • John has the flu
  • If somebody has the flu then that person has
    fever.
  • We are not certain that such assertions are true.
    We believe/disbelieve them to some degree.
    Though "belief" and "evidence" are not the same
    thing, for our purposes they will be treated
    synonymously.
  • Our problem is how to associate a degree of
    belief or of disbelief with assertions
  • How do we associate beliefs with elementary
    assertions
  • How do we combine beliefs in composite assertions
    from the beliefs of the component assertions
  • What is the relation between the beliefs of
    causally connected assertions.
  • Estimates for elementary assertions are obtained
  • From Experts (subjective probability)
  • From frequencies (if given enough data)
  • It is very hard to come up with good estimates
    for beliefs. Always consider the question "What
    if the guess is bad".
  • Estimates are needed, given the belief in
    assertions A and B, for the assertions A, A B,
    A v B
  • Evidence must be combined in cases such as
  • We have a causal connection from assertion A to
    assertion B, what can we say about B if A is
    true, or, vice versa, about A if B is true
  • We have a causal connection from assertion A to
    assertions B1 and B2, what can we say about A if
    both B1 and B2 are true
  • We have a causal connection from assertion A1 to
    B and a causal connection from A2 to B, what can
    we say about B when both A1 and A2 are true.

4
Probabilistic methods of reasoning and learning
  • Probabilistic neural networks
  • Bayesian networks
  • Markov models and chains
  • Support Vector and Kernel Machines (SVM)
  • Genetic algorithms (evolution learning)

5
  • Bayes Law
  • P(a,b) P(ab) P(b) P(ba) P(a)
  • Joint probability of a and b probability of b
    times the probability of a given b

6
Bayesian learning
7
Bayesian learning (2)
8
Bayesian learning (3)
9
Bayesian learning (4)
10
Bayes theorem
P(Aj B) posterior probability of event Aj at
condition of event B, P(B Aj)
likelihood, P(B) evidence Bayes theorem is
only valid if we know all the conditional
probabilities relating to the evidence in
question. This makes it hard to apply the theorem
in practical AI applications
11
Bayesian Network
  • A Bayesian Network is a directed acyclic graph
  • A graph where the directions are links which
    indicate dependencies that exist between nodes
    (variables).
  • Nodes represent propositions about events or
    events themselves.
  • Conditional probabilities quantify the strength
    of dependencies.
  • Consider the following example
  • The probability, that my car won't start.
  • If my car won't start then it is likely that
  • The battery is flat or
  • The staring motor is broken.
  • In order to decide whether to fix the car myself
    or send it to the garage I make the following
    decision
  • If the headlights do not work then the battery is
    likely to be flat so i fix it myself.
  • If the starting motor is defective then send car
    to garage.
  • If battery and starting motor both gone send car
    to garage.

12
A simple Bayesian network
13
Kinds of relations between variables in Bayesian
nets
  • a) Sequence, influence may be distribute from A
    to C and back while value of B is unknown
  • b) Divergence, influence may be distributed on
    childes of A while A is unknown
  • c) Convergence, about A nothing unknown except
    that may be obtained through its parents

14
Reasoning in Bayesian nets
  • Probabilities in links obey standard conditional
    probability axioms.
  • Therefore follow links in reaching hypothesis and
    update beliefs accordingly.
  • A few broad classes of algorithms have bee used
    to help with this
  • Pearls's message passing method.
  • Clique triangulation.
  • Stochastic methods.
  • Basically they all take advantage of clusters in
    the network and use their limits on the influence
    to constrain the search through net.
  • They also ensure that probabilities are updated
    correctly.
  • Since information is local information can be
    readily added and deleted with minimum effect on
    the whole network. ONLY affected nodes need
    updating.

15
Synthesis of Bayes network based on a priory
information
  • Describe task in terms of probabilities of values
    of goal variables
  • Select concept space of task, determine variables
    corresponding to goal variables, describe
    possible values of ones
  • Determine a priori probabilities of values of
    variables
  • Describe causal relations and node (variables) as
    graph
  • For every node determine condition probabilities
    of value of variable at different combinations of
    values of variables-parents

16
Applications of Bayes networks
  • Medical diagnostic systems
  • PathFinder (1990) for diagnostics of illness of
    lymphatic glands,
  • Space and military applications
  • Vista (NASA) is used for selection of needed
    information for diagnostic display from
    telemetric information in real time,
  • Netica (Australia) for defence of territory from
    sea
  • Computers and software
  • For control of agents-helpers in MS Office
  • Image processing
  • Extract of 3-dimensional scene from 2-dimensional
    images
  • Finance and economy
  • Estimation of risks and prediction of yield of
    portfolio

17
(No Transcript)
18
(No Transcript)
19
(No Transcript)
20
(No Transcript)
21
Hidden Markov Model for recognition of speech
P1,1
P2,2
P3,3
P3(.)
P1(.)
P2(.)
P1,2
P2,3
P3,4
22
Lexical HMMs
  • Create compound HMM for each lexical entry by
    concatenating the phones making up the
    pronunciation
  • example of HMM for lab (following speech for
    crossword triphone)
  • Multiple pronunciations can be weighted by
    likelihood into compound HMM for a word
  • (Tri)phone models are independent parts of word
    models

phone l a
b
triphone ch-la l-ab
a-b
Write a Comment
User Comments (0)
About PowerShow.com