Learning Models of Relational Stochastic Processes - PowerPoint PPT Presentation

About This Presentation
Title:

Learning Models of Relational Stochastic Processes

Description:

Learning Models of Relational Stochastic ... Current and Future Work Current Work Programming by Demonstration using Dynamic First Order Probabilistic ... – PowerPoint PPT presentation

Number of Views:81
Avg rating:3.0/5.0
Slides: 37
Provided by: sang73
Category:

less

Transcript and Presenter's Notes

Title: Learning Models of Relational Stochastic Processes


1
Learning Models of Relational Stochastic Processes
  • Sumit Sanghai

2
Motivation
  • Features of real-world domains
  • Multiple classes, objects, relations

3
Motivation
  • Features of real-world domains
  • Multiple classes, objects, relations
  • Uncertainty

4
Motivation
  • Features of real-world domains
  • Multiple classes, objects, relations
  • Uncertainty
  • Changes with time

P5
P2
P6
P1
P3
V1
P4
A3
A1
A2
5
Relational Stochastic Processes
  • Features
  • Multiple classes, objects, relations
  • Uncertainty
  • Change over time
  • Examples Social networks, molecular biology,
    user activity modeling, web, plan recognition,
  • Growth inherent or due to explicit actions
  • Most large datasets are gathered over time
  • Explore dependencies over time
  • Predict future

6
Manufacturing Process
7
Manufacturing Process
8
Manufacturing Process
9
Manufacturing Process
Paint(A, blue)
Bolt(B, C)
10
Why are they different?
  • Modeling object, relationships modification,
    creation and deletion
  • Modeling actions (preconditions/effects),
    activities, plans
  • Cant just throw time into the mix
  • Summarizing object information
  • Learning can be made easier by concentrating on
    temporal dependencies
  • Sophisticated inference techniques like particle
    filtering may be applicable

11
Outline
  • Background Dynamic Bayes Nets
  • Dynamic Probabilistic Relational Models
  • Inference in DPRMs
  • Learning with Dynamic Markov Logic Nets
  • Future Work

12
Dynamic Bayesian Networks
  • DBNs model change in uncertain variables over
    time
  • Each time slice consists of state/observation
    variables
  • Bayesian network models dependency of current on
    previous time slice(s)
  • At each node a conditional model (CPT, logistic
    regression, etc.)

13
Inference and learning in DBNs
  • Inference
  • All techniques from BNs are used
  • Special techniques like Particle Filtering,
    Boyen-Koller, Factored Frontier, etc. can be used
    for state monitoring
  • Learning
  • Problem exactly similar to BNs
  • Structural EM used in case of missing data
  • Needs a fast inference algorithm

14
Particle Filtering in DBNs
  • Task State monitoring
  • Particle Filter
  • Samples represent state distribution at time t
  • Generate samples for t1 based on model
  • Reweight according to observations
  • Resample
  • Particles stay in most probable regions
  • Performs poorly in hi-dimensional spaces

15
Incorporating time in First Order Probabilistic
Models
  • Simple approach Time is one of the arguments in
    first order logic
  • Year(p100, 1996), Hot (SVM, 2004)
  • But time is special
  • World is growing in the direction of time
  • Hot (SVM, 2005) dependent on Hot (SVM, 2004)
  • Hard to discover rules that help in state
    monitoring, future prediction, etc.
  • Blowup by incorporating time explicitly
  • Special inference algorithms no longer applicable

16
Dynamic Probabilistic Relational Models
  • DPRM is a PRM replicated over time slices
  • DBN is a Bayes Net replicated over time slices
  • In a DPRM attributes for each class dependent on
    attributes of same/related class
  • Related class from current/previous time slice
  • Previous relation
  • Unrolled DPRM DBN

17
DPRMs Example
t
18
Inference in DPRMs
  • Relational uncertainty ? huge state space
  • E.g. 100 parts ? 10,000 possible attachments
  • Particle filter likely to perform poorly
  • Rao-Blackwellization ??
  • Assumptions (relaxed afterwards)
  • Uncertain reference slots do not appear in slot
    chains or as parents
  • Single-valued uncertain reference slots

19
Rao-Blackwellization in DPRMs
  • Sample propositional attributes
  • Smaller space and less error
  • Constitute the particle
  • For each uncertain reference slot and particle
    state
  • Maintain a multinomial distribution over the set
    of objects in the target class
  • Conditioned on values of propositional variables

20
RBPF A Particle
Reference slots
Propositional attributes
Bolted-To-1
Bolted-To-2
Color Size Wt
Pl1 Pl2 Pl3 Pl4 Pl5
Pl1 Pl2 Pl3 Pl4 Pl5
Pl6 Pl7 Pl8 Pl9 Pl10
Red Large 2lbs
0.1 0.1 0.2 0.1 0.5
0.3 0.2 0.2 0.1 0.2
0.25 0.3 0.1 0.25 0.1

Bracket1

21
Experimental Setup
  • Assembly Domain (AIPS98)
  • Objects Plates, Brackets, Bolts
  • Attributes Color, Size, Weight, Hole type, etc.
  • Relations Bolted-To, Welded-To
  • Propositional Actions Paint, Polish, etc.
  • Relational Actions Weld, Bolt
  • Observations
  • Fault model
  • Faults cause uncertainty
  • Actions and observations uncertain
  • Governed by global fault probability (fp)
  • Task State Monitoring

22
RBPF vs PF
23
Problems with DPRMs
  • Relationships modeled using slots
  • Slots and slot chains hard to represent and
    understand
  • Modeling ternary relationships becomes hard
  • Small subset of first-order logic (conjunctive
    expressions) used to specify dependencies
  • Independence between objects participating in
    multi-valued slots
  • Unstructured conditional model

24
Relational Dynamic Bayes Nets
  • Replace slots and attributes with predicates
    (like in MLNs)
  • Each predicate has parents which are other
    predicates
  • The conditional model is a first-order
    probability tree
  • The predicate graph is acyclic
  • A copy of the model at each time slice

25
Inference Relaxing the assumptions
  • RBPF is infeasible when assumptions relaxed
  • Observation Similar objects behave similarly
  • Sample all predicates
  • Small number of samples, but large relational
    predicate space
  • Smoothing Likelihood of a small number of
    points can tell relative likelihood of others
  • Given a particle smooth each relational predicate
    towards similar states

26
Simple Smoothing
Particle Filtering A particle
Propositional attributes
Bolted-To (Bracket_1, X)
Color Size Wt
Pl1 Pl2 Pl3 Pl4 Pl5
Pl1 Pl2 Pl3 Pl4 Pl5
Red Large 2lbs
0.1 0.1 0.2 0.1 0.5
1 0 1 1 1
after smoothing
Pl1 Pl2 Pl3 Pl4 Pl5
Pl1 Pl2 Pl3 Pl4 Pl5
0.1 0.1 0.2 0.1 0.5
0.9 0.4 0.9 0.9 0.9
27
Simple Smoothing Problems
  • Simple smoothing probability of an object pair
    depends upon values of all other object pairs of
    the relation
  • E.g. P( Bolt(Br_1,Pl_1) ) depends on Bolt(Br_i,
    Pl_j) for all i and j.
  • Solution Make an object pair depend more upon
    similar pairs
  • Similarity given by properties of the objects

28
Abstraction Lattice Smoothing
  • Abstraction represents a set of similar object
    pairs.
  • Bolt(Br1, Pl1)
  • Bolt(red, large)
  • Bolt(,)
  • Abstraction Lattice a hierarchy of abstractions
  • Each abstraction has a coefficient

29
Abstraction Lattice an example
30
Abstraction Lattice Smoothing
  • P(Bolt(B1, P1)) w1 Ppf (Bolt(B1, P1)
  • w2 Ppf(Bolt(red, large))
  • w3 Ppf(Bolt(,))
  • Joint distributions are estimated using
    relational kernel density estimation
  • Kernel K(x, xi) gives distance between the state
    and the particle
  • Distance measured using abstractions

31
Abstraction Smoothing vs PF
32
Learning with DMLNs
  • Task Can MLN learning be directly applied to
    learn time-based models?
  • Domains
  • Predicting author, topic distribution in
    High-Energy Theoretical Physics papers from
    KDDCup 2003
  • Learning action models of manufacturing assembly
    processes

33
Learning with DMLNs
  • DMLNs MLNs Time predicates
  • R(x,y) -gt R(x,y,t), Succ(11, 10), Gt(10,5)
  • Now directly apply MLN structure learning
    algorithm (Stanley and Pedro)
  • To make it work
  • Use templates to model Markovian assumption
  • Restrict number of predicates per clause
  • Add background knowledge

34
Physics dataset
35
Manufacturing Assembly
36
Current and Future Work
  • Current Work
  • Programming by Demonstration using Dynamic First
    Order Probabilistic Models
  • Future Work
  • Learning object creation models
  • Learning in presence of missing data
  • Modeling hierarchies (very useful for fast
    inference)
  • Applying abstraction smoothing to static
    relational models
Write a Comment
User Comments (0)
About PowerShow.com