Mining Frequent Patterns and Association Rules - PowerPoint PPT Presentation

1 / 93
About This Presentation
Title:

Mining Frequent Patterns and Association Rules

Description:

... g., 98% of people who purchase tires and auto accessories also get automotive services done ... E.g., small sales (sum 100) trigger big buys (sum 1,000) ... – PowerPoint PPT presentation

Number of Views:194
Avg rating:3.0/5.0
Slides: 94
Provided by: jiaw189
Category:

less

Transcript and Presenter's Notes

Title: Mining Frequent Patterns and Association Rules


1
Mining Frequent Patterns and Association Rules
  • CS 536 Data Mining
  • These slides are adapted from J. Han and M.
    Kambers book slides (http//www-faculty.cs.uiuc.e
    du/hanj/bk2/)

2
What Is Frequent Pattern Analysis?
  • Frequent pattern a pattern (a set of items,
    subsequences, substructures, etc.) that occurs
    frequently in a data set
  • First proposed by Agrawal, Imielinski, and Swami
    AIS93 in the context of frequent itemsets and
    association rule mining
  • Motivation Finding inherent regularities in data
  • What products were often purchased together?
    Beer and diapers?!
  • What are the subsequent purchases after buying a
    PC?
  • What kinds of DNA are sensitive to this new drug?
  • Can we automatically classify web documents?
  • Applications
  • Basket data analysis, cross-marketing, catalog
    design, sale campaign analysis, Web log (click
    stream) analysis, and DNA sequence analysis.

3
Why Is Freq. Pattern Mining Important?
  • Discloses an intrinsic and important property of
    data sets
  • Forms the foundation for many essential data
    mining tasks
  • Association, correlation, and causality analysis
  • Sequential, structural (e.g., sub-graph) patterns
  • Pattern analysis in spatiotemporal, multimedia,
    time-series, and stream data
  • Classification associative classification
  • Cluster analysis frequent pattern-based
    clustering
  • Data warehousing iceberg cube and cube-gradient
  • Semantic data compression fascicles
  • Broad applications

4
Basic Concepts Frequent Patterns and Association
Rules
  • Itemset X x1, , xk
  • Find all the rules X ? Y with minimum support and
    confidence
  • support, s, probability that a transaction
    contains X ? Y
  • confidence, c, conditional probability that a
    transaction having X also contains Y

Transaction-id Items bought
10 A, B, D
20 A, C, D
30 A, D, E
40 B, E, F
50 B, C, D, E, F
Let supmin 50, confmin 50 Freq. Pat.
A3, B3, D4, E3, AD3 Association rules A ?
D (60, 100) D ? A (60, 75)
5
Closed Patterns and Max-Patterns
  • A long pattern contains a combinatorial number of
    sub-patterns, e.g., a1, , a100 contains (1001)
    (1002) (110000) 2100 1 1.271030
    sub-patterns!
  • Solution Mine closed patterns and max-patterns
    instead
  • An itemset X is closed if X is frequent and there
    exists no super-pattern Y ? X, with the same
    support as X (proposed by Pasquier, et al. _at_
    ICDT99)
  • An itemset X is a max-pattern if X is frequent
    and there exists no frequent super-pattern Y ? X
    (proposed by Bayardo _at_ SIGMOD98)
  • Closed pattern is a lossless compression of freq.
    patterns
  • Reducing the of patterns and rules

6
What Is Association Mining?
  • Association rule mining
  • Finding frequent patterns, associations,
    correlations, or causal structures among sets of
    items or objects in transaction databases,
    relational databases, and other information
    repositories.
  • Applications
  • Basket data analysis, cross-marketing, catalog
    design, loss-leader analysis, clustering,
    classification, etc.
  • Examples.
  • Rule form Body Head support, confidence.
  • buys(x, diapers) buys(x, lotion) 0.5,
    60
  • major(x, CS) takes(x, DB) grade(x, A)
    1, 75

7
(No Transcript)
8
Association Rule Basic Concepts
  • Given (1) database of transactions, (2) each
    transaction is a list of items (purchased by a
    customer in a visit)
  • Find all rules that correlate the presence of
    one set of items with that of another set of
    items
  • E.g., 98 of people who purchase tires and auto
    accessories also get automotive services done
  • Applications
  • ? Maintenance Agreement (What the store
    should do to boost Maintenance Agreement sales)
  • Home Electronics ? (What other products
    should the store stocks up?)
  • Attached mailing in direct marketing
  • Detecting ping-ponging of patients, faulty
    collisions

9
Rule Measures Support and Confidence
Customer buys both
  • Find all the rules X Y ? Z with minimum
    confidence and support
  • support, s, probability that a transaction
    contains X ? Y ? Z
  • confidence, c, conditional probability that a
    transaction having X ? Y also contains Z

Customer buys diaper
Customer buys lotion
  • Let minimum support 50, and minimum confidence
    50, we have
  • A ? C (50, 66.6)
  • C ? A (50, 100)

10
Association Rule Mining A Road Map
  • Boolean vs. quantitative associations (Based on
    the types of values handled)
  • buys(x, SQLServer) buys(x, DMBook)
    buys(x, DBMiner) 0.2, 60
  • age(x, 30..39) income(x, 42..48K)
    buys(x, PC) 1, 75
  • Single dimension vs. multiple dimensional
    associations (see ex. Above)
  • Single level vs. multiple-level analysis
  • What brands of lotions are associated with what
    brands of diapers?
  • Various extensions
  • Correlation, causality analysis
  • Association does not necessarily imply
    correlation or causality
  • Maxpatterns and closed itemsets
  • Constraints enforced
  • E.g., small sales (sum lt 100) trigger big buys
    (sum gt 1,000)?

11
Mining Association Rules
  • Association rule mining
  • Mining single-dimensional Boolean association
    rules from transactional databases
  • Mining multilevel association rules from
    transactional databases
  • Mining multidimensional association rules from
    transactional databases and data warehouse
  • From association mining to correlation analysis
  • Constraint-based association mining
  • Summary

12
Mining Association RulesAn Example
Min. support 50 Min. confidence 50
  • For rule A ? C
  • support support(A ?C) 50
  • confidence support(A ?C)/support(A) 66.6
  • The Apriori principle
  • Any subset of a frequent itemset must be frequent

13
Mining Frequent Itemsets the Key Step
  • Find the frequent itemsets the sets of items
    that have minimum support
  • A subset of a frequent itemset must also be a
    frequent itemset
  • i.e., if AB is a frequent itemset, both A and
    B should be a frequent itemset
  • Iteratively find frequent itemsets with
    cardinality from 1 to k (k-itemset)
  • Use the frequent itemsets to generate association
    rules.

14
The Apriori Algorithm
  • Join Step Ck is generated by joining Lk-1with
    itself
  • Prune Step Any (k-1)-itemset that is not
    frequent cannot be a subset of a frequent
    k-itemset
  • Pseudo-code
  • Ck Candidate itemset of size k
  • Lk frequent itemset of size k
  • L1 frequent items
  • for (k 1 Lk !? k) do begin
  • Ck1 candidates generated from Lk
  • for each transaction t in database do
  • increment the count of all candidates in
    Ck1 that are
    contained in t
  • Lk1 candidates in Ck1 with min_support
  • end
  • return ?k Lk

15
The Apriori Algorithm Example
Database D
L1
C1
Scan D
C2
C2
L2
Scan D
C3
L3
Scan D
16
How to Generate Candidates?
  • Suppose the items in Lk-1 are listed in an order
  • Step 1 self-joining Lk-1
  • insert into Ck
  • select p.item1 , p.item2 ,, p.itemk-2 ,
    q.itemk-1
  • from Lk-1 p, Lk-1 q
  • where p.item1q.item1 ,, p.itemk-2q.itemk-2 ,
    p.itemk-1 lt q.itemk-1
  • Step 2 pruning
  • forall itemsets c in Ck do
  • forall (k-1)-subsets s of c do
  • if (s is not in Lk-1) then delete c from Ck

17
How to Count Supports of Candidates?
  • Why counting supports of candidates a problem?
  • The total number of candidates can be very huge
  • One transaction may contain many candidates
  • Method
  • Candidate itemsets are stored in a hash-tree
  • Leaf node of hash-tree contains a list of
    itemsets and counts
  • Interior node contains a hash table
  • Subset function finds all the candidates
    contained in a transaction

18
Example of Generating Candidates
  • L3abc, abd, acd, ace, bcd
  • Self-joining L3L3
  • abcd from abc and abd
  • acde from acd and ace
  • Pruning
  • acde is removed because ade is not in L3
  • C4abcd

19
Example Transaction DB
20
Example Finding Frequent Patterns (1)
21
Example Finding Frequent Patterns (2)
22
Challenges of Frequent Pattern Mining
  • Challenges
  • Multiple scans of transaction database
  • Huge number of candidates
  • Tedious workload of support counting for
    candidates
  • Improving Apriori general ideas
  • Reduce passes of transaction database scans
  • Shrink number of candidates
  • Facilitate support counting of candidates

23
Partition Scan Database Only Twice
  • Any itemset that is potentially frequent in DB
    must be frequent in at least one of the
    partitions of DB
  • Scan 1 partition database and find local
    frequent patterns
  • Scan 2 consolidate global frequent patterns
  • A. Savasere, E. Omiecinski, and S. Navathe. An
    efficient algorithm for mining association in
    large databases. In VLDB95

24
DHP Reduce the Number of Candidates
  • A k-itemset whose corresponding hashing bucket
    count is below the threshold cannot be frequent
  • Candidates a, b, c, d, e
  • Hash entries ab, ad, ae bd, be, de
  • Frequent 1-itemset a, b, d, e
  • ab is not a candidate 2-itemset if the sum of
    count of ab, ad, ae is below support threshold
  • J. Park, M. Chen, and P. Yu. An effective
    hash-based algorithm for mining association
    rules. In SIGMOD95

25
Sampling for Frequent Patterns
  • Select a sample of original database, mine
    frequent patterns within sample using Apriori
  • Scan database once to verify frequent itemsets
    found in sample, only borders of closure of
    frequent patterns are checked
  • Example check abcd instead of ab, ac, , etc.
  • Scan database again to find missed frequent
    patterns
  • H. Toivonen. Sampling large databases for
    association rules. In VLDB96

26
DIC Reduce Number of Scans
ABCD
  • Once both A and D are determined frequent, the
    counting of AD begins
  • Once all length-2 subsets of BCD are determined
    frequent, the counting of BCD begins

ABC
ABD
ACD
BCD
AB
AC
BC
AD
BD
CD
Transactions
1-itemsets
B
C
D
A
2-itemsets
Apriori


Itemset lattice
1-itemsets
2-items
S. Brin R. Motwani, J. Ullman, and S. Tsur.
Dynamic itemset counting and implication rules
for market basket data. In SIGMOD97
3-items
DIC
27
CHARM Mining by Exploring Vertical Data Format
  • Vertical format t(AB) T11, T25,
  • tid-list list of trans.-ids containing an
    itemset
  • Deriving closed patterns based on vertical
    intersections
  • t(X) t(Y) X and Y always happen together
  • t(X) ? t(Y) transaction having X always has Y
  • Using diffset to accelerate mining
  • Only keep track of differences of tids
  • t(X) T1, T2, T3, t(XY) T1, T3
  • Diffset (XY, X) T2
  • Eclat/MaxEclat (Zaki et al. _at_KDD97), VIPER(P.
    Shenoy et al._at_SIGMOD00), CHARM (Zaki
    Hsiao_at_SDM02)

28
Is Apriori Fast Enough? Performance Bottlenecks
  • The core of the Apriori algorithm
  • Use frequent (k 1)-itemsets to generate
    candidate frequent k-itemsets
  • Use database scan and pattern matching to collect
    counts for the candidate itemsets
  • The bottleneck of Apriori candidate generation
  • Huge candidate sets
  • 104 frequent 1-itemset will generate 107
    candidate 2-itemsets
  • To discover a frequent pattern of size 100, e.g.,
    a1, a2, , a100, one needs to generate 2100 ?
    1030 candidates.
  • Multiple scans of database
  • Needs n or (n 1 ) scans, n is the length of the
    longest pattern

29
Mining Frequent Patterns Without Candidate
Generation
  • Compress a large database into a compact,
    Frequent-Pattern tree (FP-tree) structure
  • highly condensed, but complete for frequent
    pattern mining
  • avoid costly database scans
  • Develop an efficient, FP-tree-based frequent
    pattern mining method
  • A divide-and-conquer methodology decompose
    mining tasks into smaller ones
  • Avoid candidate generation sub-database test
    only!

30
Construct FP-tree from a Transaction DB
TID Items bought (ordered) frequent
items 100 f, a, c, d, g, i, m, p f, c, a, m,
p 200 a, b, c, f, l, m, o f, c, a, b,
m 300 b, f, h, j, o f, b 400 b, c, k,
s, p c, b, p 500 a, f, c, e, l, p, m,
n f, c, a, m, p
min_support 0.5
  • Steps
  • Scan DB once, find frequent 1-itemset (single
    item pattern)
  • Order frequent items in frequency descending
    order
  • Scan DB again, construct FP-tree

31
Benefits of the FP-tree Structure
  • Completeness
  • never breaks a long pattern of any transaction
  • preserves complete information for frequent
    pattern mining
  • Compactness
  • reduce irrelevant informationinfrequent items
    are gone
  • frequency descending ordering more frequent
    items are more likely to be shared
  • never be larger than the original database (if
    not count node-links and counts)
  • Example For Connect-4 DB, compression ratio
    could be over 100

32
Mining Frequent Patterns Using FP-tree
  • General idea (divide-and-conquer)
  • Recursively grow frequent pattern path using the
    FP-tree
  • Method
  • For each item, construct its conditional
    pattern-base, and then its conditional FP-tree
  • Repeat the process on each newly created
    conditional FP-tree
  • Until the resulting FP-tree is empty, or it
    contains only one path (single path will generate
    all the combinations of its sub-paths, each of
    which is a frequent pattern)

33
Major Steps to Mine FP-tree
  • Construct conditional pattern base for each node
    in the FP-tree
  • Construct conditional FP-tree from each
    conditional pattern-base
  • Recursively mine conditional FP-trees and grow
    frequent patterns obtained so far
  • If the conditional FP-tree contains a single
    path, simply enumerate all the patterns

34
Step 1 From FP-tree to Conditional Pattern Base
  • Starting at the frequent header table in the
    FP-tree
  • Traverse the FP-tree by following the link of
    each frequent item
  • Accumulate all of transformed prefix paths of
    that item to form a conditional pattern base

Conditional pattern bases item cond. pattern
base c f3 a fc3 b fca1, f1, c1 m fca2,
fcab1 p fcam2, cb1
35
Properties of FP-tree for Conditional Pattern
Base Construction
  • Node-link property
  • For any frequent item ai, all the possible
    frequent patterns that contain ai can be obtained
    by following ai's node-links, starting from ai's
    head in the FP-tree header
  • Prefix path property
  • To calculate the frequent patterns for a node ai
    in a path P, only the prefix sub-path of ai in P
    need to be accumulated, and its frequency count
    should carry the same count as node ai.

36
Step 2 Construct Conditional FP-tree
  • For each pattern-base
  • Accumulate the count for each item in the base
  • Construct the FP-tree for the frequent items of
    the pattern base

m-conditional pattern base fca2, fcab1

Header Table Item frequency head
f 4 c 4 a 3 b 3 m 3 p 3
f4
c1
All frequent patterns concerning m m, fm, cm,
am, fcm, fam, cam, fcam
b1
b1
c3
?
?
p1
a3
b1
m2
p2
m1
37
Mining Frequent Patterns by Creating Conditional
Pattern-Bases
38
Step 3 Recursively mine the conditional FP-tree
Cond. pattern base of am (fc3)

Cond. pattern base of cm (f3)
f3
cm-conditional FP-tree

Cond. pattern base of cam (f3)
f3
cam-conditional FP-tree
39
Single FP-tree Path Generation
  • Suppose an FP-tree T has a single path P
  • The complete set of frequent pattern of T can be
    generated by enumeration of all the combinations
    of the sub-paths of P


All frequent patterns concerning m m, fm, cm,
am, fcm, fam, cam, fcam
f3
?
c3
a3
m-conditional FP-tree
40
Principles of Frequent Pattern Growth
  • Pattern growth property
  • Let ? be a frequent itemset in DB, B be ?'s
    conditional pattern base, and ? be an itemset in
    B. Then ? ? ? is a frequent itemset in DB iff ?
    is frequent in B.
  • abcdef is a frequent pattern, if and only if
  • abcde is a frequent pattern, and
  • f is frequent in the set of transactions
    containing abcde

41
Why Is Frequent Pattern Growth Fast?
  • Our performance study shows
  • FP-growth is an order of magnitude faster than
    Apriori, and is also faster than tree-projection
  • Reasoning
  • No candidate generation, no candidate test
  • Use compact data structure
  • Eliminate repeated database scan
  • Basic operation is counting and FP-tree building

42
FP-growth vs. Apriori Scalability With the
Support Threshold
Data set T25I20D10K
43
FP-growth vs. Tree-Projection Scalability with
Support Threshold
Data set T25I20D100K
44
Presentation of Association Rules (Table Form )
45
Visualization of Association Rule Using Plane
Graph
46
Visualization of Association Rule Using Rule Graph
47
Mining Association Rules in Large Databases
  • Mining multilevel association rules from
    transactional databases
  • Mining multidimensional association rules from
    transactional databases and data warehouse
  • From association mining to correlation analysis
  • Constraint-based association mining
  • Summary

48
Multiple-Level Association Rules
  • Items often form hierarchy.
  • Items at the lower level are expected to have
    lower support.
  • Rules regarding itemsets at
  • appropriate levels could be quite useful.
  • Transaction database can be encoded based on
    dimensions and levels
  • We can explore shared multi-level mining

49
Mining Multi-Level Associations
  • A top_down, progressive deepening approach
  • First find high-level strong rules
  • milk bread
    20, 60.
  • Then find their lower-level weaker rules
  • 2 milk wheat
    bread 6, 50.
  • Variations at mining multiple-level association
    rules.
  • Level-crossed association rules
  • 2 milk Wonder wheat bread
  • Association rules with multiple, alternative
    hierarchies
  • 2 milk Wonder bread

50
Multi-level Association Uniform Support vs.
Reduced Support
  • Uniform Support the same minimum support for all
    levels
  • One minimum support threshold. No need to
    examine itemsets containing any item whose
    ancestors do not have minimum support.
  • Lower level items do not occur as frequently.
    If support threshold
  • too high ? miss low level associations
  • too low ? generate too many high level
    associations
  • Reduced Support reduced minimum support at lower
    levels
  • There are 4 search strategies
  • Level-by-level independent
  • Level-cross filtering by k-itemset
  • Level-cross filtering by single item
  • Controlled level-cross filtering by single item

51
Uniform Support
Multi-level mining with uniform support
Milk support 10
Level 1 min_sup 5
2 Milk support 6
Skim Milk support 4
Level 2 min_sup 5
Back
52
Reduced Support
Multi-level mining with reduced support
Level 1 min_sup 5
Milk support 10
2 Milk support 6
Skim Milk support 4
Level 2 min_sup 3
Back
53
Multi-level Association Redundancy Filtering
  • Some rules may be redundant due to ancestor
    relationships between items.
  • Example
  • milk ? wheat bread support 8, confidence
    70
  • 2 milk ? wheat bread support 2, confidence
    72
  • We say the first rule is an ancestor of the
    second rule.
  • A rule is redundant if its support is close to
    the expected value, based on the rules
    ancestor.

54
Multi-Level Mining Progressive Deepening
  • A top-down, progressive deepening approach
  • First mine high-level frequent items
  • milk (15), bread
    (10)
  • Then mine their lower-level weaker frequent
    itemsets
  • 2 milk (5),
    wheat bread (4)
  • Different min_support threshold across
    multi-levels lead to different algorithms
  • If adopting the same min_support across
    multi-levels
  • then toss t if any of ts ancestors is
    infrequent.
  • If adopting reduced min_support at lower levels
  • then examine only those descendents whose
    ancestors support is frequent/non-negligible.

55
Progressive Refinement of Data Mining Quality
  • Why progressive refinement?
  • Mining operator can be expensive or cheap, fine
    or rough
  • Trade speed with quality step-by-step
    refinement.
  • Superset coverage property
  • Preserve all the positive answersallow a
    positive false test but not a false negative
    test.
  • Two- or multi-step mining
  • First apply rough/cheap operator (superset
    coverage)
  • Then apply expensive algorithm on a substantially
    reduced candidate set (Koperski Han, SSD95).

56
Mining Association Rules in Large Databases
  • Mining multilevel association rules from
    transactional databases
  • Mining multidimensional association rules from
    transactional databases and data warehouse
  • From association mining to correlation analysis
  • Constraint-based association mining
  • Summary

57
Multi-Dimensional Association Concepts
  • Single-dimensional rules
  • buys(X, milk) ? buys(X, bread)
  • Multi-dimensional rules ? 2 dimensions or
    predicates
  • Inter-dimension association rules (no repeated
    predicates)
  • age(X,19-25) ? occupation(X,student) ?
    buys(X,coke)
  • hybrid-dimension association rules (repeated
    predicates)
  • age(X,19-25) ? buys(X, popcorn) ? buys(X,
    coke)
  • Categorical Attributes
  • finite number of possible values, no ordering
    among values
  • Quantitative Attributes
  • numeric, implicit ordering among values

58
Techniques for Mining MD Associations
  • Search for frequent k-predicate set
  • Example age, occupation, buys is a 3-predicate
    set.
  • Techniques can be categorized by how age are
    treated.
  • 1. Using static discretization of quantitative
    attributes
  • Quantitative attributes are statically
    discretized by using predefined concept
    hierarchies.
  • 2. Quantitative association rules
  • Quantitative attributes are dynamically
    discretized into binsbased on the distribution
    of the data.
  • 3. Distance-based association rules
  • This is a dynamic discretization process that
    considers the distance between data points.

59
Static Discretization of Quantitative Attributes
  • Discretized prior to mining using concept
    hierarchy.
  • Numeric values are replaced by ranges.
  • In relational database, finding all frequent
    k-predicate sets will require k or k1 table
    scans.
  • Data cube is well suited for mining.
  • The cells of an n-dimensional
  • cuboid correspond to the
  • predicate sets.
  • Mining from data cubescan be much faster.

60
Quantitative Association Rules
  • Numeric attributes are dynamically discretized
  • Such that the confidence or compactness of the
    rules mined is maximized.
  • 2-D quantitative association rules Aquan1 ?
    Aquan2 ? Acat
  • Cluster adjacent
  • association rules
  • to form general
  • rules using a 2-D
  • grid.
  • Example

age(X,30-34) ? income(X,24K - 48K) ?
buys(X,high resolution TV)
61
ARCS (Association Rule Clustering System)
  • How does ARCS work?
  • 1. Binning
  • 2. Find frequent predicateset
  • 3. Clustering
  • 4. Optimize

62
Limitations of ARCS
  • Only quantitative attributes on LHS of rules.
  • Only 2 attributes on LHS. (2D limitation)
  • An alternative to ARCS
  • Non-grid-based
  • equi-depth binning
  • clustering based on a measure of partial
    completeness.
  • Mining Quantitative Association Rules in Large
    Relational Tables by R. Srikant and R. Agrawal.

63
Mining Distance-based Association Rules
  • Binning methods do not capture the semantics of
    interval data
  • Distance-based partitioning, more meaningful
    discretization considering
  • density/number of points in an interval
  • closeness of points in an interval

64
Clusters and Distance Measurements
  • SX is a set of N tuples t1, t2, , tN ,
    projected on the attribute set X
  • The diameter of SX
  • distxdistance metric, e.g. Euclidean distance or
    Manhattan

65
Clusters and Distance Measurements(Cont.)
  • The diameter, d, assesses the density of a
    cluster CX , where
  • Finding clusters and distance-based rules
  • the density threshold, d0 , replaces the notion
    of support
  • modified version of the BIRCH clustering
    algorithm

66
Mining Association Rules in Large Databases
  • Mining multilevel association rules from
    transactional databases
  • Mining multidimensional association rules from
    transactional databases and data warehouse
  • From association mining to correlation analysis
  • Constraint-based association mining
  • Summary

67
Interestingness Measurements
  • Objective measures
  • Two popular measurements
  • support and
  • confidence
  • Subjective measures (Silberschatz Tuzhilin,
    KDD95)
  • A rule (pattern) is interesting if
  • it is unexpected (surprising to the user) and/or
  • actionable (the user can do something with it)

68
Criticism to Support and Confidence
  • Example 1 (Aggarwal Yu, PODS98)
  • Among 5000 students
  • 3000 play basketball
  • 3750 eat cereal
  • 2000 both play basket ball and eat cereal
  • play basketball ? eat cereal 40, 66.7 is
    misleading because the overall percentage of
    students eating cereal is 75 which is higher
    than 66.7.
  • play basketball ? not eat cereal 20, 33.3 is
    far more accurate, although with lower support
    and confidence

69
Criticism to Support and Confidence (Cont.)
  • Example 2
  • X and Y positively correlated,
  • X and Z, negatively related
  • support and confidence of
  • XgtZ dominates
  • We need a measure of dependent or correlated
    events
  • P(BA)/P(B) is also called the lift of rule A gt B

70
Other Interestingness Measures Interest
  • Interest (correlation, lift)
  • taking both P(A) and P(B) in consideration
  • P(AB)P(B)P(A), if A and B are independent
    events
  • A and B negatively correlated, if the value is
    less than 1 otherwise A and B positively
    correlated

71
Mining Association Rules in Large Databases
  • Mining multilevel association rules from
    transactional databases
  • Mining multidimensional association rules from
    transactional databases and data warehouse
  • From association mining to correlation analysis
  • Constraint-based association mining
  • Summary

72
Constraint-Based Mining
  • Interactive, exploratory mining giga-bytes of
    data?
  • Could it be real? Making good use of
    constraints!
  • What kinds of constraints can be used in mining?
  • Knowledge type constraint classification,
    association, etc.
  • Data constraint SQL-like queries
  • Find product pairs sold together in Vancouver in
    Dec.98.
  • Dimension/level constraints
  • in relevance to region, price, brand, customer
    category.
  • Rule constraints
  • small sales (price lt 10) triggers big sales
    (sum gt 200).
  • Interestingness constraints
  • strong rules (min_support ? 3, min_confidence ?
    60).

73
Rule Constraints in Association Mining
  • Two kind of rule constraints
  • Rule form constraints meta-rule guided mining.
  • P(x, y) Q(x, w) takes(x, database
    systems).
  • Rule (content) constraint constraint-based query
    optimization (Ng, et al., SIGMOD98).
  • sum(LHS) lt 100 min(LHS) gt 20 count(LHS) gt 3
    sum(RHS) gt 1000
  • 1-variable vs. 2-variable constraints
    (Lakshmanan, et al. SIGMOD99)
  • 1-var A constraint confining only one side (L/R)
    of the rule, e.g., as shown above.
  • 2-var A constraint confining both sides (L and
    R).
  • sum(LHS) lt min(RHS) max(RHS) lt 5 sum(LHS)

74
Constrain-Based Association Query
  • Database (1) trans (TID, Itemset ), (2)
    itemInfo (Item, Type, Price)
  • A constrained asso. query (CAQ) is in the form of
    (S1, S2 )C ,
  • where C is a set of constraints on S1, S2
    including frequency constraint
  • A classification of (single-variable)
    constraints
  • Class constraint S ? A. e.g. S ? Item
  • Domain constraint
  • S? v, ? ? ?, ?, ?, ?, ?, ? . e.g. S.Price lt
    100
  • v? S, ? is ? or ?. e.g. snacks ? S.Type
  • V? S, or S? V, ? ? ?, ?, ?, ?, ?
  • e.g. snacks, sodas ? S.Type
  • Aggregation constraint agg(S) ? v, where agg is
    in min, max, sum, count, avg, and ? ? ?, ?,
    ?, ?, ?, ? .
  • e.g. count(S1.Type) ? 1 , avg(S2.Price) ? 100

75
Constrained Association Query Optimization Problem
  • Given a CAQ (S1, S2) C , the algorithm
    should be
  • sound It only finds frequent sets that satisfy
    the given constraints C
  • complete All frequent sets satisfy the given
    constraints C are found
  • A naïve solution
  • Apply Apriori for finding all frequent sets, and
    then to test them for constraint satisfaction one
    by one.
  • Better approach
  • Comprehensive analysis of the properties of
    constraints and try to push them as deeply as
    possible inside the frequent set computation.

76
Anti-monotone and Monotone Constraints
  • A constraint Ca is anti-monotone iff. for any
    pattern S not satisfying Ca, none of the
    super-patterns of S can satisfy Ca
  • A constraint Cm is monotone iff. for any pattern
    S satisfying Cm, every super-pattern of S also
    satisfies it

77
Anti-Monotonicity in Constraint Pushing
TDB (min_sup2)
  • Anti-monotonicity
  • When an intemset S violates the constraint, so
    does any of its superset
  • sum(S.Price) ? v is anti-monotone
  • sum(S.Price) ? v is not anti-monotone
  • Example. C range(S.profit) ? 15 is anti-monotone
  • Itemset ab violates C
  • So does every superset of ab

TID Transaction
10 a, b, c, d, f
20 b, c, d, f, g, h
30 a, c, d, e, f
40 c, e, f, g
Item Profit
a 40
b 0
c -20
d 10
e -30
f 30
g 20
h -10
78
Monotonicity for Constraint Pushing
TDB (min_sup2)
  • Monotonicity
  • When an intemset S satisfies the constraint, so
    does any of its superset
  • sum(S.Price) ? v is monotone
  • min(S.Price) ? v is monotone
  • Example. C range(S.profit) ? 15
  • Itemset ab satisfies C
  • So does every superset of ab

TID Transaction
10 a, b, c, d, f
20 b, c, d, f, g, h
30 a, c, d, e, f
40 c, e, f, g
Item Profit
a 40
b 0
c -20
d 10
e -30
f 30
g 20
h -10
79
Succinct Constraint
  • A subset of item Is is a succinct set, if it can
    be expressed as ?p(I) for some selection
    predicate p, where ? is a selection operator
  • A constraint Cs is succinct provided that all
    sets satifying the constraint are a succinct set.

80
Succinctness
  • Succinctness
  • Given A1, the set of items satisfying a
    succinctness constraint C, then any set S
    satisfying C is based on A1 , i.e., S contains a
    subset belonging to A1
  • Idea Without looking at the transaction
    database, whether an itemset S satisfies
    constraint C can be determined based on the
    selection of items
  • min(S.Price) ? v is succinct
  • sum(S.Price) ? v is not succinct
  • Optimization If C is succinct, C is pre-counting
    pushable

81
The Apriori Algorithm Example
Database D
L1
C1
Scan D
C2
C2
L2
Scan D
C3
L3
Scan D
82
Naïve Algorithm Apriori Constraint
Database D
L1
C1
Scan D
C2
C2
L2
Scan D
C3
L3
Constraint SumS.price lt 5
Scan D
83
The Constrained Apriori Algorithm Push an
Anti-monotone Constraint Deep
Database D
L1
C1
Scan D
C2
C2
L2
Scan D
C3
L3
Constraint SumS.price lt 5
Scan D
84
The Constrained Apriori Algorithm Push a
Succinct Constraint Deep
Database D
L1
C1
Scan D
C2
C2
L2
Scan D
not immediately to be used
C3
L3
Constraint minS.price lt 1
Scan D
85
Converting Tough Constraints
TDB (min_sup2)
TID Transaction
10 a, b, c, d, f
20 b, c, d, f, g, h
30 a, c, d, e, f
40 c, e, f, g
  • Convert tough constraints into anti-monotone or
    monotone by properly ordering items
  • Examine C avg(S.profit) ? 25
  • Order items in value-descending order
  • lta, f, g, d, b, h, c, egt
  • If an itemset afb violates C
  • So does afbh, afb
  • It becomes anti-monotone!

Item Profit
a 40
b 0
c -20
d 10
e -30
f 30
g 20
h -10
86
Convertible Constraint
  • Suppose all items in patterns are listed in a
    total order R
  • A constraint C is convertible anti-monotone iff a
    pattern S satisfying the constraint implies that
    each suffix of S w.r.t. R also satisfies C
  • A constraint C is convertible monotone iff a
    pattern S satisfying the constraint implies that
    each pattern of which S is a suffix w.r.t. R also
    satisfies C

87
What Constraints Are Convertible?
Constraint Convertible anti-monotone Convertible monotone Strongly convertible
avg(S) ? , ? v Yes Yes Yes
median(S) ? , ? v Yes Yes Yes
sum(S) ? v (items could be of any value, v ? 0) Yes No No
sum(S) ? v (items could be of any value, v ? 0) No Yes No
sum(S) ? v (items could be of any value, v ? 0) No Yes No
sum(S) ? v (items could be of any value, v ? 0) Yes No No

88
Constraint-Based MiningA General Picture
Constraint Antimonotone Monotone Succinct
v ? S no yes yes
S ? V no yes yes
S ? V yes no yes
min(S) ? v no yes yes
min(S) ? v yes no yes
max(S) ? v yes no yes
max(S) ? v no yes yes
count(S) ? v yes no weakly
count(S) ? v no yes weakly
sum(S) ? v ( a ? S, a ? 0 ) yes no no
sum(S) ? v ( a ? S, a ? 0 ) no yes no
range(S) ? v yes no no
range(S) ? v no yes no
avg(S) ? v, ? ? ?, ?, ? convertible convertible no
support(S) ? ? yes no no
support(S) ? ? no yes no
89
Relationships Among Categories of Constraints
Succinctness
Anti-monotonicity
Monotonicity
Convertible constraints
Inconvertible constraints
90
Summary
  • Association rule mining
  • probably the most significant contribution from
    the database community in KDD
  • A large number of papers have been published
  • Many interesting issues have been explored
  • An interesting research direction
  • Association analysis in other types of data
    spatial data, multimedia data, time series data,
    etc.

91
References
  • R. Agarwal, C. Aggarwal, and V. V. V. Prasad. A
    tree projection algorithm for generation of
    frequent itemsets. In Journal of Parallel and
    Distributed Computing (Special Issue on High
    Performance Data Mining), 2000.
  • R. Agrawal, T. Imielinski, and A. Swami. Mining
    association rules between sets of items in large
    databases. SIGMOD'93, 207-216, Washington, D.C.
  • R. Agrawal and R. Srikant. Fast algorithms for
    mining association rules. VLDB'94 487-499,
    Santiago, Chile.
  • R. Agrawal and R. Srikant. Mining sequential
    patterns. ICDE'95, 3-14, Taipei, Taiwan.
  • R. J. Bayardo. Efficiently mining long patterns
    from databases. SIGMOD'98, 85-93, Seattle,
    Washington.
  • S. Brin, R. Motwani, and C. Silverstein. Beyond
    market basket Generalizing association rules to
    correlations. SIGMOD'97, 265-276, Tucson,
    Arizona.
  • S. Brin, R. Motwani, J. D. Ullman, and S. Tsur.
    Dynamic itemset counting and implication rules
    for market basket analysis. SIGMOD'97, 255-264,
    Tucson, Arizona, May 1997.
  • K. Beyer and R. Ramakrishnan. Bottom-up
    computation of sparse and iceberg cubes.
    SIGMOD'99, 359-370, Philadelphia, PA, June 1999.
  • D.W. Cheung, J. Han, V. Ng, and C.Y. Wong.
    Maintenance of discovered association rules in
    large databases An incremental updating
    technique. ICDE'96, 106-114, New Orleans, LA.
  • M. Fang, N. Shivakumar, H. Garcia-Molina, R.
    Motwani, and J. D. Ullman. Computing iceberg
    queries efficiently. VLDB'98, 299-310, New York,
    NY, Aug. 1998.

92
References (2)
  • G. Grahne, L. Lakshmanan, and X. Wang. Efficient
    mining of constrained correlated sets. ICDE'00,
    512-521, San Diego, CA, Feb. 2000.
  • Y. Fu and J. Han. Meta-rule-guided mining of
    association rules in relational databases.
    KDOOD'95, 39-46, Singapore, Dec. 1995.
  • T. Fukuda, Y. Morimoto, S. Morishita, and T.
    Tokuyama. Data mining using two-dimensional
    optimized association rules Scheme, algorithms,
    and visualization. SIGMOD'96, 13-23, Montreal,
    Canada.
  • E.-H. Han, G. Karypis, and V. Kumar. Scalable
    parallel data mining for association rules.
    SIGMOD'97, 277-288, Tucson, Arizona.
  • J. Han, G. Dong, and Y. Yin. Efficient mining of
    partial periodic patterns in time series
    database. ICDE'99, Sydney, Australia.
  • J. Han and Y. Fu. Discovery of multiple-level
    association rules from large databases. VLDB'95,
    420-431, Zurich, Switzerland.
  • J. Han, J. Pei, and Y. Yin. Mining frequent
    patterns without candidate generation. SIGMOD'00,
    1-12, Dallas, TX, May 2000.
  • T. Imielinski and H. Mannila. A database
    perspective on knowledge discovery.
    Communications of ACM, 3958-64, 1996.
  • M. Kamber, J. Han, and J. Y. Chiang.
    Metarule-guided mining of multi-dimensional
    association rules using data cubes. KDD'97,
    207-210, Newport Beach, California.
  • M. Klemettinen, H. Mannila, P. Ronkainen, H.
    Toivonen, and A.I. Verkamo. Finding interesting
    rules from large sets of discovered association
    rules. CIKM'94, 401-408, Gaithersburg, Maryland.

93
References (3)
  • F. Korn, A. Labrinidis, Y. Kotidis, and C.
    Faloutsos. Ratio rules A new paradigm for fast,
    quantifiable data mining. VLDB'98, 582-593, New
    York, NY.
  • B. Lent, A. Swami, and J. Widom. Clustering
    association rules. ICDE'97, 220-231, Birmingham,
    England.
  • H. Lu, J. Han, and L. Feng. Stock movement and
    n-dimensional inter-transaction association
    rules. SIGMOD Workshop on Research Issues on
    Data Mining and Knowledge Discovery (DMKD'98),
    121-127, Seattle, Washington.
  • H. Mannila, H. Toivonen, and A. I. Verkamo.
    Efficient algorithms for discovering association
    rules. KDD'94, 181-192, Seattle, WA, July 1994.
  • H. Mannila, H Toivonen, and A. I. Verkamo.
    Discovery of frequent episodes in event
    sequences. Data Mining and Knowledge Discovery,
    1259-289, 1997.
  • R. Meo, G. Psaila, and S. Ceri. A new SQL-like
    operator for mining association rules. VLDB'96,
    122-133, Bombay, India.
  • R.J. Miller and Y. Yang. Association rules over
    interval data. SIGMOD'97, 452-461, Tucson,
    Arizona.
  • R. Ng, L. V. S. Lakshmanan, J. Han, and A. Pang.
    Exploratory mining and pruning optimizations of
    constrained associations rules. SIGMOD'98, 13-24,
    Seattle, Washington.
  • N. Pasquier, Y. Bastide, R. Taouil, and L.
    Lakhal. Discovering frequent closed itemsets for
    association rules. ICDT'99, 398-416, Jerusalem,
    Israel, Jan. 1999.

94
References (4)
  • J.S. Park, M.S. Chen, and P.S. Yu. An effective
    hash-based algorithm for mining association
    rules. SIGMOD'95, 175-186, San Jose, CA, May
    1995.
  • J. Pei, J. Han, and R. Mao. CLOSET An Efficient
    Algorithm for Mining Frequent Closed Itemsets.
    DMKD'00, Dallas, TX, 11-20, May 2000.
  • J. Pei and J. Han. Can We Push More Constraints
    into Frequent Pattern Mining? KDD'00. Boston,
    MA. Aug. 2000.
  • G. Piatetsky-Shapiro. Discovery, analysis, and
    presentation of strong rules. In G.
    Piatetsky-Shapiro and W. J. Frawley, editors,
    Knowledge Discovery in Databases, 229-238.
    AAAI/MIT Press, 1991.
  • B. Ozden, S. Ramaswamy, and A. Silberschatz.
    Cyclic association rules. ICDE'98, 412-421,
    Orlando, FL.
  • J.S. Park, M.S. Chen, and P.S. Yu. An effective
    hash-based algorithm for mining association
    rules. SIGMOD'95, 175-186, San Jose, CA.
  • S. Ramaswamy, S. Mahajan, and A. Silberschatz. On
    the discovery of interesting patterns in
    association rules. VLDB'98, 368-379, New York,
    NY..
  • S. Sarawagi, S. Thomas, and R. Agrawal.
    Integrating association rule mining with
    relational database systems Alternatives and
    implications. SIGMOD'98, 343-354, Seattle, WA.
  • A. Savasere, E. Omiecinski, and S. Navathe. An
    efficient algorithm for mining association rules
    in large databases. VLDB'95, 432-443, Zurich,
    Switzerland.
  • A. Savasere, E. Omiecinski, and S. Navathe.
    Mining for strong negative associations in a
    large database of customer transactions. ICDE'98,
    494-502, Orlando, FL, Feb. 1998.

95
References (5)
  • C. Silverstein, S. Brin, R. Motwani, and J.
    Ullman. Scalable techniques for mining causal
    structures. VLDB'98, 594-605, New York, NY.
  • R. Srikant and R. Agrawal. Mining generalized
    association rules. VLDB'95, 407-419, Zurich,
    Switzerland, Sept. 1995.
  • R. Srikant and R. Agrawal. Mining quantitative
    association rules in large relational tables.
    SIGMOD'96, 1-12, Montreal, Canada.
  • R. Srikant, Q. Vu, and R. Agrawal. Mining
    association rules with item constraints. KDD'97,
    67-73, Newport Beach, California.
  • H. Toivonen. Sampling large databases for
    association rules. VLDB'96, 134-145, Bombay,
    India, Sept. 1996.
  • D. Tsur, J. D. Ullman, S. Abitboul, C. Clifton,
    R. Motwani, and S. Nestorov. Query flocks A
    generalization of association-rule mining.
    SIGMOD'98, 1-12, Seattle, Washington.
  • K. Yoda, T. Fukuda, Y. Morimoto, S. Morishita,
    and T. Tokuyama. Computing optimized rectilinear
    regions for association rules. KDD'97, 96-103,
    Newport Beach, CA, Aug. 1997.
  • M. J. Zaki, S. Parthasarathy, M. Ogihara, and W.
    Li. Parallel algorithm for discovery of
    association rules. Data Mining and Knowledge
    Discovery, 1343-374, 1997.
  • M. Zaki. Generating Non-Redundant Association
    Rules. KDD'00. Boston, MA. Aug. 2000.
  • O. R. Zaiane, J. Han, and H. Zhu. Mining
    Recurrent Items in Multimedia with Progressive
    Resolution Refinement. ICDE'00, 461-470, San
    Diego, CA, Feb. 2000.
Write a Comment
User Comments (0)
About PowerShow.com