Title: Prof NB Venkateswarlu
1- Prof NB Venkateswarlu
- Head, IT, GVPCOE
- Visakhapatnam
- venkat_ritch_at_yahoo.com
- www.ritchcenter.com/nbv
2First Let me say Hearty Welcome to you All
3- Also, let me
- congrachulate
- Chairman,
- Secretary/Correspondent
4- Principal,
- Prof. Ravindra Babu
- Vice-Principal
5- and other Organizers for planning for such a
nice workshop with excellent themes.
6My Talk
- Feature Extraction/ Selection
7A Typical Image Processing System contains
Image Acquisition
Image Pre-Processing
Image En-hancement
Image Seg-mentation
Image Featu-re Extraction
Image Class-fication
Image Unde-rstanding
8Two Aspects of Feature Extraction
- Extracting useful features from images or any
other measurements.
9- Identifying Transformed Variables which are
functions of original variables and having some
charcateristics.
10Feature Selection
-
- Selecting Important Variables is Feature
Selection
11- Some Features Used in I.P Applications
12- Shape based
- Contour based
- Area based
- Transform based
- Projections
- Signature
- Problem specific
13Perimeter, length etc. First Convex hull is
extracted
14Skeletons
15(No Transcript)
16Averaged Radial density
17Radial Basis functions
18Rose Plots
19Chain Codes
20Crack code - 32330300
21Signature
22Bending Energy
23Chord Distribution
24Fourier Descriptors
25Structure
26Splines
27Horizontal and vertical projections
28Elongatedness
29Convex Hull
30Compactness
31RGB, R ,G and B bands
32(No Transcript)
33Classification/Pattern Recognition
- Statistical
- Syntactical Linguistic
- Discriminant function
- Fuzzy
- Neural
- Hybrid
34Dimensionality Reduction
- Feature selection (i.e., attribute subset
selection) - Select a minimum set of features such that the
probability distribution of different classes
given the values for those features is as close
as possible to the original distribution given
the values of all features - reduce of patterns in the patterns, easier to
understand - Heuristic methods (due to exponential of
choices) - step-wise forward selection
- step-wise backward elimination
- combining forward selection and backward
elimination - decision-tree induction
35Example of Decision Tree Induction
- Initial attribute set
- A1, A2, A3, A4, A5, A6
36Heuristic Feature Selection Methods
- There are 2d possible sub-features of d features
- Several heuristic feature selection methods
- Best single features under the feature
independence assumption choose by significance
tests. - Best step-wise feature selection
- The best single-feature is picked first
- Then next best feature condition to the first,
... - Step-wise feature elimination
- Repeatedly eliminate the worst feature
- Best combined feature selection and elimination
- Optimal branch and bound
- Use feature elimination and backtracking
37Why do We need?
- A classifier performance depends on
- No of features
- Feature distinguishability
- No of groups
- Groups characteristics in multidimensional space.
- Needed response time
- Memory requirements
38Feature Extraction Methods
- We will find transformed variables which are
functions of original variables. - A good example Though we may conduct tests in
more than test (K-D), finally grading is done
based on total marks (1-D)
39Principal Component Analysis
- Given N data vectors from k-dimensions, find c lt
k orthogonal vectors that can be best used to
represent data - The original data set is reduced to one
consisting of N data vectors on c principal
components (reduced dimensions) - Each data vector is a linear combination of the c
principal component vectors - Works for numeric data only
- Used when the number of dimensions is large
40Principal Component Analysis
41Principal Component Analysis
- Aimed at finding new co-ordinate system which has
some characteristics. - M4.5 4.25
- Cov Matrix 2.57 1.86
- 1.86 6.21
- Eigen Values 6.99, 1.79
- Eigen Vectors 0.387 0.922
- -0.922 0.387
42(No Transcript)
43However in some cases it is not possible to have
PCA working.
44Canonical Analysis
45- Unlike PCA which takes global mean and
covariance, this takes between the group and
within the group covariance matrix and the
calculates canonical axes.
46Standard Deviation A Simple Indicator
47Feature Selection Group Separability Indices
48(No Transcript)
49(No Transcript)
50(No Transcript)
51(No Transcript)
52Feature Selection Through Clustering
53(No Transcript)
54Selecting From 4 variables
55Multi-Layer Perceptron
56Network Pruning and Rule Extraction
- Network pruning
- Fully connected network will be hard to
articulate - N input nodes, h hidden nodes and m output nodes
lead to h(mN) weights - Pruning Remove some of the links without
affecting classification accuracy of the network - Extracting rules from a trained network
- Discretize activation values replace individual
activation value by the cluster average
maintaining the network accuracy - Enumerate the output from the discretized
activation values to find rules between
activation value and output - Find the relationship between the input and
activation value - Combine the above two to have rules relating the
output to input
57Neural Networks for Feature Extraction
58Self-organizing feature maps (SOMs)
- Clustering is also performed by having several
units competing for the current object - The unit whose weight vector is closest to the
current object wins - The winner and its neighbors learn by having
their weights adjusted - SOMs are believed to resemble processing that can
occur in the brain - Useful for visualizing high-dimensional data in
2- or 3-D space
59Other Model-Based Clustering Methods
- Neural network approaches
- Represent each cluster as an exemplar, acting as
a prototype of the cluster - New objects are distributed to the cluster whose
exemplar is the most similar according to some
dostance measure - Competitive learning
- Involves a hierarchical architecture of several
units (neurons) - Neurons compete in a winner-takes-all fashion
for the object currently being presented
60Model-Based Clustering Methods
61SVM
- SVM constructs nonlinear decision functions by
training classifier to perform a linear
separation in some high dimensional space which
is nonlinearly related to the input space. A
Mercer kernel is used for mapping.
62(No Transcript)