Title: Invariant%20Local%20Feature%20for%20Object%20Recognition
1Invariant Local Feature for Object Recognition
- Presented by Wyman
- 2/05/2006
2Introduction
- Object Recognition
- A task of finding 3D objects from 2D images (or
even video) and classifying them into one of the
many known object types - Closely related to the success of many computer
vision applications - robotics, surveillance, registration etc.
- A difficult problem that a general and
comprehensive solution to this problem has not
been made
3Introduction
- Two main streams of approaches
- Model-Based Object Recognition
- View-Based Object Recognition
- 2D representations of the same object viewed at
different angles and distances when available - Extract features (as the representations of
object) and compare them to those in the feature
database
4Matching with Local Features
- One of the possible solution
- Matching with invariant local features
- Robust to Occlusion, clutter background
- cf. global features
- Three phases
- Detection
- Description
- Matching
Accurate, Fast
Distinctive
Invariance
5Research Direction
- Study and improve the invariant local features
- Detection, description and matching
- Study and improve object recognition / matching
using invariant local features - Area to improve
- Distinctiveness
- Invariance
- Efficiency
6Outline
- State-of-the-art techniques
- Descriptor
- Matching
- Conclusion Future Works
7Outline
- State-of-the-art techniques
- Descriptor
- Performance evaluation
- Current extension using color
- Possible way to improve Color Orientation
- Matching
- Conclusion Future Work
8Outline
- State-of-the-art techniques
- Descriptor
- Performance evaluation
- Current extension using color
- Possible way to improve Color Orientation
- Matching
- Cross-bin distance
- Performance evaluation
- Possible way to improve Aggregation of Content
- Conclusion Future Work
9Performance Evaluation of Descriptors
- We aim to compare the performance of three
state-of-the-art local feature descriptors SIFT,
PCA-SIFT and GLOH - Same experimental setup as that used in
Performance Evaluation of Local Descriptors
TPAMI 2005 - Different evaluation criterion
- Different result
- In each experiment, each descriptor describe
features from - Harris corner detector
- Harris-affine covariant detector
- Output regions that are invariant to viewpoint
change
10SIFT Scale Invariant Feature Transform
Detector Descriptor Descriptor Descriptor
Invariance Scale Rotation Illumination Viewpoint
- Descriptor overview
- Find local orientation as the dominant gradient
direction ? Rotation Invariant - Compute gradient orientation histograms of
several small windows (128 values for each point)
relative to the local orientation ? Viewpoint
Invariant - Normalize the descriptor to make it invariant to
intensity change ? Illumination
D.Lowe. Distinctive Image Features from
Scale-Invariant Keypoints. IJCV 2004
11PCA-SIFT
- Rotate feature region to dominant gradient
direction same as SIFT - Pre-compute an eigenspace for local gradient
patches of size 41x41 - 2x39x393042 elements
- Only keep 20 components
- A more compact descriptor
- Sensitive to viewpoint change
Y. K. Rahul. Pca-sift A more distinctive
representation for local image descriptors. CVPR
2004
12GLOH (Gradient location-orientation histogram)
- Different from SIFT in sampling method
- 17 log-polar location bins
- 16 orientation bins
- Analyze the 17x16272 Dimensions
- Apply PCA analysis, keep 128 components
PCA on Orientation Histogram VS PCA on Gradient
Patch
17 Log-polar location bins
C. S. Krystian Mikolajczyk. A performance
evaluation of local descriptors. TPAMI 2005
13Performance Evaluation
Scale Rotation (bark)
- Data Set
- From Visual Geometry Group
Viewpoint change (graf)
Illumination change (leuven)
Viewpoint change (wall)
Blur
Blurring (bikes)
14Performance Evaluation
- Evaluation Criteria
- Match features from first image to the second one
based on the nearest neighbor distance ratio - That is, two features are matched if first
nearest neighbor is much closer than the second
nearest neighbor - This is different from the threshold-based
criterion used in A Performance Evaluation of
Local Descriptors TPAMI 2005 - Count the number of correct matches and the
number of false matches obtained for an image
pair - The results are plotted in form of recall versus
1-precision curves
15Performance Evaluation
Viewpoint change (wall)
Scale Rotation (bark)
Illumination change (leuven)
Blurring (bikes)
16Performance Evaluation Result
Descriptor Distinctiveness Complexity Feature Size
SIFT High Medium 128
PCA-SIFT Medium Low 20
GLOH High High 128
- For accuracy ? SIFT
- For speed ? PCA-SIFT
- In large database ? ?
17Start from Scratch
- Comparison of my descriptor with SIFT
- Simply designed vs carefully designed
- Result
- SIFT is a carefully designed descriptor, it
remains robust when the degree of transformation
increases
Increasing illumination change
Increasing affine change
Increasing affine change
Increasing blur
18Extension using Color
- Weijier extends local feature descriptors with
color information, by concatenating a color
descriptor, K, to the shape descriptor, S,
according to - where B is the combined color and shape
descriptor and is a weighting parameter and
indicates that the vector is normalized.
J. van de Weijer and C. Schmid. Coloring local
feature extraction. ECCV2006.
19Proposed Extension using Color
- Problem statement
- Orientation of local feature patch are obtained
from the monochrome intensity image - Color feature patches on the right has the same
grayscale patches shown on the left. Thus, they
are assigned the same orientation histogram - If we can generate significant orientation
histogram for each of them, we can further
improve the distinctiveness of the shape
descriptor, SIFT
20Feature Matching
- Original distance metric designed for SIFT,
PCA-SIFT and GLOH is bin-to-bin Euclidean
distance - Problems
- Sensitive to quantization effects
- Sensitive to distortion problems due to
deformation, illumination change and noise
21Feature Matching Diffusion Distance
- Haibin Ling proposed a new distance metric for
histogram-based descriptor called diffusion
distance - Summing value in all layers of the distance
pyramid with exponentially decreasing size
Gaussian Blur In 3 directions 3D case
Gaussian Blur In 1 direction 1D case
H. Ling and K. Okada. Diffusion distance for
histogram comparison. CVPR06.
22Feature Matching Performance Evaluation
- Same setup as the previous experiment
- Recall vs 1-prevision curve for image pair with
affine transformation
23Feature Matching Performance Evaluation
Data set. The synthetic deformation data set from
Haibin Ling
Images in the data set and the evaluation method
needs to be improved
24Proposed Extension
- Robust aggregation of the histogram, such as
average orientation direction and center of mass
of derivatives, can be also used in comparison - Diffusion distance can be viewed as a form of
comparison using the aggregate information - Its aggregation of histogram bins is obtained by
repeatedly convolving the histogram with Gaussian
kernels - Summation of the distance between each
aggregation pair of two histograms gives the
diffusion distance
Histogram A
Histogram B
128 bins
128 bins
64 bins
64 bins
32 bins
32 bins
Aggregation 1. Average of gradient magnitude
over location bins 2. Bin reduction in
orientation bins
25Conclusion and Future Work
- Presented
- Result of performance evaluation of some
state-of-the-art descriptors and feature matching
distance metric - Possible way to improve the description and
matching step - TODO
- Incorporate color information into local features
- Improve features distinctiveness
- Design a distance metric for comparing SIFT
features histogram - Invariant to deformation (like diffusion
distance) - Improve features distinctiveness
26Q A
27Models of Image Change
- Geometry
- Rotation
- Similarity (rotation uniform scale)
- Affine (scale dependent on direction)valid for
orthographic camera, locally planar object - Photometry
- Affine intensity change (I ? a I b)
28Image Alignment
- Many applications
- 3D reconstruction, motion tracking, indexing and
database retrieval, robot navigation - Image alignment for building panorama
29Image Alignment
- Detect features in both images
30Image Alignment
- Detect features in both images
- Find corresponding pairs
31Image Alignment
- Detect features in both images
- Find corresponding pairs
- Use these pairs to align images
32Difficulties
- Problem 1
- Detect the same point independently in both images
no chance to match!
We need a repeatable detector
33Difficulties
- Problem 2
- For each point correctly recognize the
corresponding one
?
We need a reliable and distinctive descriptor
34Difficulties
- Problem 3
- Image transformation may exist in the two images
- Change in scale, rotation, illumination and
viewpoint
?
We need an invariant local feature descriptor