Flexible Metric NN Classification - PowerPoint PPT Presentation

About This Presentation
Title:

Flexible Metric NN Classification

Description:

k-NN assigns an unknown object to the most common class of its k nearest neighbors ... M1th order statistic. Results on Artificial Data. Results on Real Data ... – PowerPoint PPT presentation

Number of Views:17
Avg rating:3.0/5.0
Slides: 14
Provided by: davidm123
Category:

less

Transcript and Presenter's Notes

Title: Flexible Metric NN Classification


1
Flexible Metric NN Classification
  • based on Friedman (1995)
  • David Madigan

2
Nearest-Neighbor Methods
  • k-NN assigns an unknown object to the most common
    class of its k nearest neighbors
  • Choice of k? (bias-variance tradeoff again)
  • Choice of metric?
  • Need all the training to be present to classify a
    new point (lazy methods)
  • Surprisingly strong asymptotic results (e.g. no
    decision rule is more than twice as accurate as
    1-NN)

3
(No Transcript)
4
Suppose a Regression Surface Looks like this
want this
not this
Flexible-metric NN Methods try to capture this
idea
5
FMNN
  • Predictors may not all be equally relevant for
    classifying a new object
  • Furthermore, this differential relevance may
    depend on the location of the new object
  • FMNN attempts to model this phenomenon

6
Local Relevance
  • Consider an arbitrary function f on Rp
  • If no values of x are known, have
  • Suppose xiz, then

7
Local Relevance cont.
  • The improvement in squared error provided by
    knowing xi is
  • I2i(z) reflects the importance of the ith
    variable on the variation of f(x) at xiz

8
Local Relevance cont.
  • Now consider an arbitrary point z(z1,,zp)
  • The relative importance of xi to the variation of
    f at xz is
  • R2i(z)0 when f(x) is independent of xi at z
  • R2i(z)1 when f(x) depends only on xi at z

9
Estimation
  • Recall

10
On To Classification
  • For J-class classification have yj, j1,,J
    output variables, yj e 0,1, S yj1.
  • Can compute
  • Technical point need to weight the observations
    to rectify unequal variances

11
The Machete
  • Start with all data points R0
  • Compute
  • Then
  • Continue until Ri contains K points

M1th order statistic
12
Results on Artificial Data
13
Results on Real Data
Write a Comment
User Comments (0)
About PowerShow.com