Introduction to Truth Maintenance Systems - PowerPoint PPT Presentation

About This Presentation
Title:

Introduction to Truth Maintenance Systems

Description:

of the lack of information, or lack of time to derive the conclusion. ... When the domain evolves, the new domain state may be ... – PowerPoint PPT presentation

Number of Views:1757
Avg rating:3.0/5.0
Slides: 15
Provided by: csC5
Learn more at: https://cs.ccsu.edu
Category:

less

Transcript and Presenter's Notes

Title: Introduction to Truth Maintenance Systems


1
Introduction to Truth Maintenance Systems
  • A Truth Maintenance System (TMS) is a PS module
    responsible for
  • Enforcing logical relations among beliefs.
  • Generating explanations for conclusions.
  • Finding solutions to search problems
  • Supporting default reasoning.
  • Identifying causes for failure and recover from
    inconsistencies.
  • The TMS / IE relationship is the following

Justifications, assumptions
Inference Engine
TMS
Beliefs, contradictions
Problem Solver
2
  • 1. Enforcement of logical relations
    (constrains) among beliefs.
  • Every AI problem which is not completely
    specified requires search. Search
  • utilizes assumptions, which may eventually
    change.Changing assumptions
  • requires updating consequences of beliefs.
    Re-derivation of those
  • consequences is most often not desirable,
    therefore we need a mechanism to
  • maintain and update relations among beliefs.
  • Example If (cs-501) and (math-218) then
    (cs-570).
  • If (cs-570) and
    (CIT-core-completed) then (TMS-related-capstone).
  • If (TMS-related-capstone) then
    (AI-experience).
  • The following are relations among beliefs
    following from these statements
  • (AI-experience) if
    (TMS-related-capstone).
  • (TMS-related-capstone) if
    (cs-570), (CIT-core-completed).
  • etc.
  • Beliefs can be viewed as propositional variables,
    and a TMS can be viewed as

3
  • 2. Generation of explanations.
  • Solving problems is what PSs do. However, often
    solutions are not enough -
  • the PS is expected to provide an explanation for
    the proposed solution so that
  • the user can identify the cause of a problem if
    something goes wrong. To
  • provide explanations, a TMS uses cached
    inferences.
  • The fundamental assumption behind this idea
    is that caching inferences
  • once is more beneficial than running
    inference rules that have generated
  • these inferences more than once.
  • Example Q Shall I have an AI experience after
    completing the CIT program?
  • A Yes, because of the TMS
    related capstone.
  • Q What do I need to take a TMS
    related capstone?
  • A CS-570 and completed core.
  • Note There are different types of TMSs that
    provide different ways of explaining
  • conclusions (JTMS vs ATMS). In this example,
    explaining conclusions in terms
  • of their immediate predecessors works much better.

4
  • 3. Finding solutions to search problems.
  • Consider the following graph
  • B
  • A D
  • C E
  • Assume you want to color the nodes so that every
    node is red, or green, or
  • yellow, and adjacent nodes are of different
    colors. Let "1" means "red", "2"
  • means "green", and "3" means "yellow". Then, the
    following set of constraints
  • describe this problem
  • A1 or A2 or A3 not (A1 and B1)
    not (A3 and C3) not (D2 and E2)
  • B1 or B2 or B3 not (A2 and B2)
    not (B1 and D1) not (D3 and E3)
  • C1 or C2 or C2 not (A3 and B3)
    not (B2 and D2) not (C1 and E1)
  • D1 or D2 or D3 not (A1 and C1)
    not (B3 and D3) not (C2 and E2)
  • E1 or E2 or E2 not (A2 and C2)
    not (D1 and E1) not (C3 and E3)

5
  • To find a solution that satisfies all of the
    constraints, we can use search
  • A is red A is
    green A is yellow
  • B is red B is green B is yellow
  • C is red C is green C is yellow
  • D is red D is green D is yellow
  • E is red E is green E is yellow

6
  • 4. Default reasoning and TMS
  • Many real-world problems cannot be completely
    specified. That is, the PS must
  • make conclusions based on incomplete information.
    Typically the assumption
  • under which such conclusions are drawn is that X
    is true unless there is an
  • evidence to the contrary. This is known as the
    Closed-World Assumption
  • (CWA). Notice that the CWA helps us limit the
    underlying search space by
  • assuming only a certain choice and ignoring the
    others. The reasoning scheme
  • that utilizes this assumption is called default
    (or non-monotonic) reasoning.
  • Example Consider the following knowledge base
  • Bird(tom) and not Abnormal(tom) ?
    Can_fly(tom)
  • Penguin(tom) ? Abnormal(tom)
  • Ostrich(tom) ? Abnormal(tom)
  • Bird(tom)
  • -------------------------------------------
    --
  • Under the CWA, we assume not Abnormal(tom)
    (because there is no
  • evidence that Tom is abnormal). Therefore,
    we can derive can_fly(tom).

7
  • 5. Identifying causes for failures and recovering
    from inconsistencies.
  • Inconsistencies among beliefs in the KB are
    always possible, especially if the PS
  • makes its conclusions based on insufficient
    information. The most common
  • reasons for inconsistencies or other failures
    are the following
  • -- Wrong data. Example Outside temperature is
    320 degrees.
  • -- Impossible constraints. Example (Big-house
    and Cheap-house and
  • Nice-house).
  • -- Non-monotonic inference. PS is forced to
    jump to a conclusion, because
  • of the lack of information, or lack of time to
    derive the conclusion.
  • -- Contradictions due to inconsistent data,
    conclusions contradicting the
  • existing data, or inconsistent assumptions.
  • -- Dynamic data. When the domain evolves, the new
    domain state may be
  • considerably different from the previous
    domain state, and inferences
  • made in the previous state may no longer be
    valid.
  • Cashed dependences among beliefs that TMS
    maintains help identify the reason
  • for an inconsistency, and a mechanism, called
    dependency-directed

8
How the TMS and the IE communicate?
  • The PS works with
  • assertions (facts, beliefs, conclusions,
    hypotheses)
  • inference rules
  • procedures.
  • Each one of these is assigned a TMS node.
  • Example
  • N1 (rule (student ?x)
  • (assert (and (underpaid
    ?x) (overworked ?x))))
  • N2 (student Bob)
  • Note that the IE and the TMS treat nodes
    differently. Given N1 and N2, the IE
  • can infer
  • N3 (and (underpaid Bob) (overworked
    Bob))
  • This is possible because the IE threats nodes as
    logical formulas, while the TMS
  • treats nodes as propositional variables.

9
TMS nodes
  • Different types of TMSs support different types
    of nodes. Here are the basic ones
  • Premise nodes. These are always true.
  • Contradiction nodes. These are always false.
  • Assumption nodes. These are nodes, which the IE
    wants to believe no matter whether or not they
    are supported by the existing evidence.
  • (Regular) nodes. These are nodes which are
    believed only if there is a valid reason for
    that.
  • Each node has a label associated with it. The
    contents and the structure of the
  • label depends on the type of TMS. In the simplest
    case, it may only indicate
  • whether a node is believed (IN) or not believed
    (OUT).
  • Nodes are complex data structures, where
    different node properties are stored.
  • Labels are just one of those properties. Other
    properties are node type (premise,
  • assumption, etc.), node support (justifications,
    antecedents), node consequences,
  • etc.

10
TMS justifications
  • Once a new node, N3, is created by the IE, it can
    be reported to the TMS
  • together with the fact that it follows from N1,
    N2 and MP. This is recorded in the
  • following form, called the justification
  • (N3 Modus-Ponens N2 N1)
  • Here N3 is called the consequent, Modus-Ponens is
    the informant, N1 and
  • N2 are the antecedents of the justification. That
    is, justifications record
  • relations among beliefs (N1, N2 and N3 in this
    case), and therefore can be
  • used for explaining consequents and identifying
    causes for inconsistencies.
  • The general format of justifications is the
    following
  • (ltconsequentgt ltinformantgt . ltantecedentsgt)

11
TMS dependency networks
  • Nodes and justifications form a dependency
    network. Here is an example
  • network
  • Node
  • Justifications for Node
  • one of them is selected to
  • be Node's "support", i.e Node's
    consequences these are justifications
  • the reason for Node to for other
    nodes of the network, for which Node
  • be believed ("IN"). is an
    antecedent.
  • See figures 6.8 and 6.9 for examples.

12
TMS / IE interaction
  • Responsibilities of the IE
  • Adds assertions and justifications.
  • Makes premises and assumptions,
  • Retracts assumptions.
  • Provides advise on handling contradictions
  • Responsibilities of the TMS
  • Cashes beliefs and consequences and maintains
    labels.
  • Detects contradictions.
  • Performed belief revision.
  • Generates explanations.

13
Propositional specification of a TMS
  • As we have already seen, TMS nodes are
    propositional variables. Therefore,
  • we can view TMS justifications as propositional
    formulas (implications) of the
  • form
  • N1 N2 Ni ? Nj
  • Here N1, N2, , Ni, Nj are positive literals,
    therefore this implication is a Horn
  • formula.
  • A TMS can be viewed as a collection
    of Horn formulas.
  • There exist polynomial time inference procedures
    for Horn formulas knowledge
  • bases. For example, forward chaining -- by just
    applying MP, we can derive all
  • formulas that logically follow from the KB. This
    makes it possible for a TMS to
  • answer a variety of queries about the current set
    of nodes and justifications.
  • The most fundamental query is whether a node
    logically follows from a given
  • TMS state.

14
Families of TMSs
  • There are several families of TMSs, which differ
    in the representation scheme
  • they use and the functionality they support
  • Justification-based TMSs. The language used is
    limited to Horn formulas.
  • Logic-based TMSs. These use a full propositional
    logic language.
  • Assumption-based TMSs. Language limited to Horn
    formulas, but several alternatives (contexts) can
    be explored at the same time.
  • Non-monotonic JTMSs. Language limited to Horn
    formulas, but allow non-monotonic justifications,
    thus making it possible to implement default
    reasoning.
  • Clause Management Systems. Their representational
    power is equivalent to LTMSs, but like ATMSs can
    support several contexts at the same time.
  • Contradiction-tolerant TMSs. Language limited to
    Horn formulas, but support non-monotonic and
    plausible reasoning and deal explicitly with
    contradictions in a single context.
Write a Comment
User Comments (0)
About PowerShow.com