Title: Subspace Clustering
1Subspace Clustering
- Ali Sekmen and Ghan S. Bhatt
- Computer Science and Mathematical Sciences
- College of Engineering
- Tennessee State University
1st Annual Workshop on Data Sciences
2Part I
- Some Linear Algebra
- Spectral Analysis
- Singular Value Decomposition
- Presenter
- Dr. Ghan S. Bhatt
3Definitions
4Range and Null Spaces
5Range and Null Spaces
6Definitions
7Eigenvalues - Eigenvectors
8Eigenvalues - Eigenvectors
9Eigenvalues - Eigenvectors
10Eigenvalues - Eigenvectors
11Symmetric Matrices
12Symmetric Matrices
13Projection on a Vector
14Projection on a Subspace
15Singular Value Decomposition
16Singular Value Decomposition
17Singular Value Decomposition
18Singular Value Decomposition
19Important Lemma
20Recall Linear Mapping
21Recall Linear Mapping
22General Matrix Norms
23An Intuitive Matrix Norm
This satisfies the general matrix norm properties
Although it is useful, it is not suitable for
large set of problems and we need another
definition of matrix norms
24Induced Matrix Norms
25Matrix p-Norm
26More on Matrix Norms
27Part II
- Subspace Segmentation Problem
- Motion Segmentation
- Principal Component Analysis
- Dimensionality Reduction
- Spectral Clustering
- Presenter
- Dr. Ali Sekmen
28Subspace Segmentation
- In many engineering and mathematics applications,
data lives in a union of low dimensional
subspaces - Motion segmentation
- Facial images of a person with the same
expression under different illumination
approximately lie on the same subspace
29Face Recognition
30Problem Statement
31Problem Statement
32Problem Statement
33What are we trying to solve?
34Example Motion Segmentation
35Motion Segmentation
Motion segmentation problem can simply be defined
as identifying independently moving rigid objects
in a video.
36Motion Segmentation
37Motion Segmentation
Z
Y
X
38Motion Segmentation
39Motion Segmentation
40Motion Segmentation
41Motion Segmentation
Y
X
42Motion Segmentation
Motion Segmentation
43Motion Segmentation
44Motion Segmentation
45Principal Component Analysis
- The goal is to reduce dimension of dataset with
minimal loss of information - We project a feature space onto a smaller
subspace that represent data well - Search for a subspace which maximizes the
variance of projected points - This is equivalent to linear least square fitting
- Minimize the sum of squared distances between
points and subspace - We find directions (components) that maximizes
variance in dataset - PCA can be done by
- Eigenvalue decomposition of a data covariance
matrix - Or SVD of a data matrix
46Least Square Approximation
47Principal Component Analysis
48Principal Component Analysis
49PCA with SVD
Coordinates w.r.t. new basis
50Principal Component Analysis
inch
cm
51Principal Component Analysis
inch cm
10 28
12 19
15 40
20 47
23 56
26 69
52Solution with SVD
53PCA Pre-Processing
inch cm
10 28
12 19
15 40
20 47
23 56
26 69
54PCA Optimization
55PCA Reduce Dimensionality
56PCA Reduce Dimensionality
57General PCA
58Spectral Clustering
- A very powerful clustering algorithm
- Easy to implement
- Outperforms traditional clustering algorithms
- Example k-means
- It is not easy to understand why it works
- Given a set of data points and some similarity
measure between all pairs of data points, we
divide data into groups - Points in the same group are similar
- Points in different groups are dissimilar
59Spectral Clustering
- Most of subspace clustering algorithms employ
spectral clustering as the last step
60Similarity
61Spectral Clustering
62Spectral Clustering
63Spectral Clustering
64Spectral Clustering Example
From Lecture Notes of Ulrike von Luxburg
65Spectral Clustering Example
From Lecture Notes of Ulrike von Luxburg
66Spectral Clustering Example
From Lecture Notes of Ulrike von Luxburg
67Spectral Clustering Example
From Lecture Notes of Ulrike von Luxburg
68Spectral Clustering Example
From Lecture Notes of Ulrike von Luxburg
69Spectral Clustering Example
70Spectral Clustering Example
71Spectral Clustering Example
72(No Transcript)