Decision Trees with Numeric Tests - PowerPoint PPT Presentation

About This Presentation
Title:

Decision Trees with Numeric Tests

Description:

Title: Data Mining lecture Author: Arno Knobbe Last modified by: Arno Knobbe Created Date: 6/4/1996 5:33:28 PM Document presentation format: Letter Paper (8.5x11 in) – PowerPoint PPT presentation

Number of Views:26
Avg rating:3.0/5.0
Slides: 19
Provided by: ArnoK1
Category:

less

Transcript and Presenter's Notes

Title: Decision Trees with Numeric Tests


1
Decision Trees with Numeric Tests
2
Industrial-strength algorithms
  • For an algorithm to be useful in a wide range of
    real-world applications it must
  • Permit numeric attributes
  • Allow missing values
  • Be robust in the presence of noise
  • Basic schemes need to be extended to fulfill
    these requirements

3
C4.5 History
  • ID3, CHAID 1960s
  • C4.5 innovations (Quinlan)
  • permit numeric attributes
  • deal sensibly with missing values
  • pruning to deal with for noisy data
  • C4.5 - one of best-known and most widely-used
    learning algorithms
  • Last research version C4.8, implemented in Weka
    as J4.8 (Java)
  • Commercial successor C5.0 (available from
    Rulequest)

4
Numeric attributes
  • Standard method binary splits
  • E.g. temp lt 45
  • Unlike nominal attributes,every attribute has
    many possible split points
  • Solution is straightforward extension
  • Evaluate info gain (or other measure)for every
    possible split point of attribute
  • Choose best split point
  • Info gain for best split point is info gain for
    attribute
  • Computationally more demanding

5
Example
  • Split on temperature attribute
  • E.g. temperature ? 71.5 yes/4, no/2 temperature
    ? 71.5 yes/5, no/3
  • Info(4,2,5,3) 6/14 info(4,2) 8/14
    info(5,3) 0.939 bits
  • Place split points halfway between values
  • Can evaluate all split points in one pass!

64 65 68 69 70 71 72 72 75 75 80 81 83 85 Yes No Yes Yes Yes No No Yes Yes Yes No Yes Yes No
6
Example
  • Split on temperature attribute

64 65 68 69 70 71 72 72 75 75 80 81 83 85 Yes No Yes Yes Yes No No Yes Yes Yes No Yes Yes No
infoGain
0
7
Speeding up
  • Entropy only needs to be evaluated between points
    of different classes (Fayyad Irani, 1992)

value class
64 65 68 69 70 71 72 72 75 75 80 81 83 85 Yes No Yes Yes Yes No No Yes Yes Yes No Yes Yes No
Potential optimal breakpoints Breakpoints
between values of the same class cannot be optimal
8
Missing as a separate value
  • Missing value denoted ? in C4.5
  • Simple idea treat missing as a separate value
  • Q When is this not appropriate?
  • A When values are missing due to different
    reasons
  • Example 1 gene expression could be missing when
    it is very high or very low
  • Example 2 field IsPregnantmissing for a male
    patient should be treated differently (no) than
    for a female patient of age 25 (unknown)

9
Missing values - advanced
  • Split instances with missing values into pieces
  • A piece going down a branch receives a weight
    proportional to the popularity of the branch
  • weights sum to 1
  • Info gain works with fractional instances
  • use sums of weights instead of counts
  • During classification, split the instance into
    pieces in the same way
  • Merge probability distribution using weights

10
Application Computer Vision 1
11
Application Computer Vision 2
  • feature extraction
  • color (RGB, hue, saturation)
  • edge, orientation
  • texture
  • XY coordinates
  • 3D information

12
Application Computer Vision 3
how grey?
below horizon?
13
Application Computer Vision 4
  • prediction

14
Application Computer Vision 4
  • inverse perspective

15
Application Computer Vision 5
  • inverse perspective
  • path planning

16
Quiz 1
  • Q If an attribute A has high info gain, does it
    always appear in a decision tree?
  • A No.
  • If it is highly correlated with another attribute
    B, and infoGain(B) gt infoGain(A), then B will
    appear in the tree, and further splitting on A
    will not be useful.

17
Quiz 2
  • Q Can an attribute appear more than once in a
    decision tree?
  • A Yes.
  • If a test is not at the root of the tree, it can
    appear in different branches.
  • Q And on a single path in the tree (from root to
    leaf)?
  • A Yes.
  • Numeric attributes can appear more than once, but
    only with very different numeric conditions.

18
Quiz 3
  • Q If an attribute A has infoGain(A)0, can it
    ever appear in a decision tree?
  • A Yes.
  • All attributes may have zero info gain.
  • info gain often changes when splitting on another
    attribute.
Write a Comment
User Comments (0)
About PowerShow.com