SlideShare a Scribd company logo
1 of 57
CS 312 Character and
Pattern Recognition – Spring 2009
Dimensionality Reduction Using PCA/LDA
Chapter 3 (Duda et al.) – Section 3.8
Thanks to Prof. Bebis
Case Studies:
Face Recognition Using Dimensionality Reduction
M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive Neuroscience, 3(1),
pp. 71-86, 1991.
D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image Retrieval", IEEE
Transactions on Pattern Analysis and Machine Intelligence, 18(8), pp. 831-836, 1996.
A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on Pattern Analysis and
Machine Intelligence, vol. 23, no. 2, pp. 228-233, 2001.
2
Dimensionality Reduction
• One approach to deal with high dimensional data is by reducing
their dimensionality.
• Project high dimensional data onto a lower dimensional sub-
space using linear or non-linear transformations.
3
Dimensionality Reduction
• Linear transformations are simple to compute and tractable.
• Classical –linear- approaches:
– Principal Component Analysis (PCA)
– Fisher Discriminant Analysis (FDA)
( )t
i i iY U X b u a= =
k x 1 k x d d x 1 (k<<d)k x 1 k x d d x 1 (k<<d)
4
Principal Component Analysis (PCA)
• Each dimensionality reduction technique finds an
appropriate transformation by satisfying certain criteria
(e.g., information loss, data discrimination, etc.)
• The goal of PCA is to reduce the dimensionality of the data
while retaining as much as possible of the variation
present in the dataset.
5
Principal Component Analysis (PCA)
1 1 2 2
1 2
ˆ ...
where , ,..., isa basein the -dimensionalsub-space(K<N)
K K
K
x b u b u b u
u u u K
= + + +
ˆx x=
1 1 2 2
1 2
...
where , ,..., isa basein theoriginal N-dimensionalspace
N N
n
x a v a v a v
v v v
= + + +
• Find a basis in a low dimensional sub-space:
− Approximate vectors by projecting them in a low dimensional
sub-space:
(1) Original space representation:
(2) Lower-dimensional sub-space representation:
• Note: if K=N, then
6
Principal Component Analysis (PCA)
• Example (K=N):
7
Principal Component Analysis (PCA)
• Information loss
− Dimensionality reduction implies information loss !!
− PCA preserves as much information as possible:
• What is the “best” lower dimensional sub-space?
The “best” low-dimensional space is centered at the sample mean
and has directions determined by the “best” eigenvectors of the
covariance matrix of the data x.
− By “best” eigenvectors we mean those corresponding to the largest
eigenvalues ( i.e., “principal components”).
− Since the covariance matrix is real and symmetric, these
eigenvectors are orthogonal and form a set of basis vectors.
(see pp. 114-117 in textbook for a proof)
ˆmin || || (reconstruction error)x x−
8
Principal Component Analysis (PCA)
• Methodology
− Suppose x1, x2, ..., xM are N x 1 vectors
9
Principal Component Analysis (PCA)
• Methodology – cont.
( )T
i ib u x x= −
10
Principal Component Analysis (PCA)
• Eigenvalue spectrum
λi
K
λN
11
Principal Component Analysis (PCA)
• Linear transformation implied by PCA
− The linear transformation RN
→ RK
that performs the dimensionality
reduction is:
12
Principal Component Analysis (PCA)
• Geometric interpretation
− PCA projects the data along the directions where the data varies the
most.
− These directions are determined by the eigenvectors of the
covariance matrix corresponding to the largest eigenvalues.
− The magnitude of the eigenvalues corresponds to the variance of
the data along the eigenvector directions.
13
Principal Component Analysis (PCA)
• How many principal components to keep?
− To choose K, you can use the following criterion:
14
Principal Component Analysis (PCA)
• What is the error due to dimensionality reduction?
• It can be shown that the average error due to
dimensionality reduction is equal to:
e
15
Principal Component Analysis (PCA)
• Standardization
− The principal components are dependent on the units used to
measure the original variables as well as on the range of values
they assume.
− We should always standardize the data prior to using PCA.
− A common standardization method is to transform all the data to
have zero mean and unit standard deviation:
16
Principal Component Analysis (PCA)
• Case Study: Eigenfaces for Face Detection/Recognition
− M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of
Cognitive Neuroscience, vol. 3, no. 1, pp. 71-86, 1991.
• Face Recognition
− The simplest approach is to think of it as a template matching
problem
− Problems arise when performing recognition in a high-
dimensional space.
− Significant improvements can be achieved by first mapping
the data into a lower dimensionality space.
− How to find this lower-dimensional space?
17
Principal Component Analysis (PCA)
• Main idea behind eigenfaces
ˆΦ −Ψ
−Ψ
average face
18
Principal Component Analysis (PCA)
• Computation of the eigenfaces
19
Principal Component Analysis (PCA)
• Computation of the eigenfaces – cont.
20
Principal Component Analysis (PCA)
• Computation of the eigenfaces – cont.
ui
21
Principal Component Analysis (PCA)
• Computation of the eigenfaces – cont.
22
Principal Component Analysis (PCA)
• Representing faces onto this basis
ˆ
iΦ
23
Principal Component Analysis (PCA)
• Eigenvalue spectrum : Defined as set of Eigenvalues.
• Spectrum is shown as Eigenvalues plotted in the graph
shown below:
λi
K
M
λN
24
Principal Component Analysis (PCA)
• Representing faces onto this basis – cont.
25
Principal Component Analysis (PCA)
• Face Recognition Using Eigenfaces
26
Principal Component Analysis (PCA)
• Face Recognition Using Eigenfaces – cont.
− The distance er is called distance within the face space (difs)
− Comment: we can use the common Euclidean distance to compute
er, however, it has been reported that the Mahalanobis distance
performs better:
27
Principal Component Analysis (PCA)
• Face Detection Using Eigenfaces
28
Principal Component Analysis (PCA)
• Face Detection Using Eigenfaces – cont.
29
Principal Component Analysis (PCA)
• Reconstruction of
faces and non-faces
30
Principal Component Analysis (PCA)
• Applications
− Face detection, tracking, and recognition
dffs
31
Principal Component Analysis (PCA)
• Problems in face recognition:
− Background (de-emphasize the outside of the face – e.g., by
multiplying the input image by a 2D Gaussian window centered on
the face)
− Lighting conditions (performance degrades with light changes)
− Scale (performance decreases quickly with changes to head size)
− multi-scale eigenspaces
− scale input image to multiple sizes
− Orientation (performance decreases but not as fast as with scale
changes)
− plane rotations can be handled
− out-of-plane rotations are more difficult to handle
32
Principal Component Analysis (PCA)
• Dataset
− 16 subjects
− 3 orientations, 3 sizes
− 3 lighting conditions, 6 resolutions (512x512 ... 16x16)
− Total number of images: 2,592
33
Principal Component Analysis (PCA)
• Experiment 1
− Used various sets of 16 images for training
− One image/person, taken under the same conditions
− Classify the rest images as one of the 16 individuals
− 7 eigenfaces were used
− No rejections (i.e., no threshold on difs)
− Performed a large number of experiments and averaged the results:
− 96% correct averaged over light variation
− 85% correct averaged over orientation variation
− 64% correct averaged over size variation
34
Principal Component Analysis (PCA)
• Experiment 2
− Considered rejections (i.e., by thresholding difs)
− There is a tradeoff between correct recognition and rejections.
− Adjusting the threshold to achieve 100% recognition accuracy
resulted in:
− 19% rejections while varying lighting
− 39% rejections while varying orientation
− 60% rejections while varying size
• Experiment 3
− Reconstruction using
partial information
35
Principal Component Analysis (PCA)
• PCA and classification
− PCA is not always an optimal dimensionality-reduction procedure
for classification purposes.
• Multiple classes and PCA
− Suppose there are C classes in the training data.
− PCA is based on the sample covariance which characterizes the
scatter of the entire data set, irrespective of class-membership.
− The projection axes chosen by PCA might not provide good
discrimination power.
36
Principal Component Analysis (PCA)
• PCA and classification (cont’d)
37
Linear Discriminant Analysis (LDA)
• What is the goal of LDA?
− Perform dimensionality reduction “while preserving as much of
the class discriminatory information as possible”.
− Seeks to find directions along which the classes are best separated.
− Takes into consideration the scatter within-classes but also the
scatter between-classes.
− More capable of distinguishing image variation due to identity from
variation due to other sources such as illumination and expression.
38
Linear Discriminant Analysis (LDA)
39
Linear Discriminant Analysis (LDA)
• Notation
(Sb has at most rank C-1)
(each sub-matrix has
rank 1 or less, i.e., outer
product of two vectors)
1 1
( )( )
iMC
T
w j i j i
i j
S x x
= =
= − −∑∑ μ μ
40
Linear Discriminant Analysis (LDA)
• Methodology
− LDA computes a transformation that maximizes the between-class
scatter while minimizing the within-class scatter:
| | | |
max max
| | | |
T
b b
T
w w
U S U S
U S U S
=
%
%
T
U=y x
products of eigenvalues !
projection matrix
,b wS S% % : scatter matrices of the projected data y
41
Linear Discriminant Analysis (LDA)
• Linear transformation implied by LDA
− The linear transformation is given by a matrix U whose columns are
the eigenvectors of the above problem (i.e., called Fisherfaces).
− The LDA solution is given by the eigenvectors of the generalized
eigenvector problem:
− Important: Since Sb has at most rank C-1, the max number of
eigenvectors with non-zero eigenvalues is C-1 (i.e., max
dimensionality of sub-space is C-1)
42
Linear Discriminant Analysis (LDA)
• Does Sw
-1
always exist?
− If Sw is non-singular, we can obtain a conventional eigenvalue
problem by writing:
− In practice, Sw is often singular since the data are image vectors
with large dimensionality while the size of the data set is much
smaller (M << N )
43
Linear Discriminant Analysis (LDA)
• Does Sw
-1
always exist? – cont.
− To alleviate this problem, we can use PCA first:
1) PCA is first applied to the data set to reduce its dimensionality.
2) LDA is then applied to find the most discriminative directions:
44
Linear Discriminant Analysis (LDA)
• Case Study: Using Discriminant Eigenfeatures for Image
Retrieval
− D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image
Retrieval", IEEE Transactions on Pattern Analysis and Machine
Intelligence, vol. 18, no. 8, pp. 831-836, 1996.
• Content-based image retrieval
− The application being studied here is query-by-example image
retrieval.
− The paper deals with the problem of selecting a good set of image
features for content-based image retrieval.
45
Linear Discriminant Analysis (LDA)
• Assumptions
− "Well-framed" images are required as input for training and query-
by-example test probes.
− Only a small variation in the size, position, and orientation of the
objects in the images is allowed.
46
Linear Discriminant Analysis (LDA)
• Some terminology
− Most Expressive Features (MEF): the features (projections)
obtained using PCA.
− Most Discriminating Features (MDF): the features (projections)
obtained using LDA.
• Numerical problems
− When computing the eigenvalues/eigenvectors of Sw
-1
SBuk = λkuk
numerically, the computations can be unstable since Sw
-1
SB is not
always symmetric.
− See paper for a way to find the eigenvalues/eigenvectors in a stable
way.
47
Linear Discriminant Analysis (LDA)
• Factors unrelated to classification
− MEF vectors show the tendency of PCA to capture major variations
in the training set such as lighting direction.
− MDF vectors discount those factors unrelated to classification.
48
Linear Discriminant Analysis (LDA)
• Clustering effect
1) Generate the set of MEFs/MDFs for each image in the training set.
2) Given an query image, compute its MEFs/MDFs using the same
procedure.
3) Find the k closest neighbors for retrieval (e.g., using Euclidean
distance).
• Methodology
49
Linear Discriminant Analysis (LDA)
• Experiments and results
− Face images
− A set of face images was used with 2 expressions, 3 lighting conditions.
− Testing was performed using a disjoint set of images:
• One image, randomly chosen, from each individual.
50
Linear Discriminant Analysis (LDA)
51
Linear Discriminant Analysis (LDA)
− Examples of correct search probes
52
Linear Discriminant Analysis (LDA)
− Example of a failed search probe
53
Linear Discriminant Analysis (LDA)
• Case Study: PCA versus LDA
− A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on
Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp. 228-
233, 2001.
• Is LDA always better than PCA?
− There has been a tendency in the computer vision community to
prefer LDA over PCA.
− This is mainly because LDA deals directly with discrimination
between classes while PCA does not pay attention to the underlying
class structure.
− Main results of this study:
(1) When the training set is small, PCA can outperform LDA.
(2) When the number of samples is large and representative for each
class, LDA outperforms PCA.
54
Linear Discriminant Analysis (LDA)
• Is LDA always better than PCA? – cont.
55
Linear Discriminant Analysis (LDA)
• Is LDA always better than PCA? – cont.
LDA is not always better when training set is small
56
Linear Discriminant Analysis (LDA)
• Is LDA always better than PCA? – cont.
LDA outperforms PCA when training set is large
END
57

More Related Content

What's hot

3.3 hierarchical methods
3.3 hierarchical methods3.3 hierarchical methods
3.3 hierarchical methodsKrish_ver2
 
Presentation on unsupervised learning
Presentation on unsupervised learning Presentation on unsupervised learning
Presentation on unsupervised learning ANKUSH PAL
 
Optimization/Gradient Descent
Optimization/Gradient DescentOptimization/Gradient Descent
Optimization/Gradient Descentkandelin
 
3.2 partitioning methods
3.2 partitioning methods3.2 partitioning methods
3.2 partitioning methodsKrish_ver2
 
07 dimensionality reduction
07 dimensionality reduction07 dimensionality reduction
07 dimensionality reductionMarco Quartulli
 
Image segmentation
Image segmentationImage segmentation
Image segmentationkhyati gupta
 
Random forest algorithm
Random forest algorithmRandom forest algorithm
Random forest algorithmRashid Ansari
 
Types of clustering and different types of clustering algorithms
Types of clustering and different types of clustering algorithmsTypes of clustering and different types of clustering algorithms
Types of clustering and different types of clustering algorithmsPrashanth Guntal
 
Unsupervised learning (clustering)
Unsupervised learning (clustering)Unsupervised learning (clustering)
Unsupervised learning (clustering)Pravinkumar Landge
 
04 Classification in Data Mining
04 Classification in Data Mining04 Classification in Data Mining
04 Classification in Data MiningValerii Klymchuk
 
Hierarchical clustering.pptx
Hierarchical clustering.pptxHierarchical clustering.pptx
Hierarchical clustering.pptxNTUConcepts1
 

What's hot (20)

Pca ppt
Pca pptPca ppt
Pca ppt
 
3. mining frequent patterns
3. mining frequent patterns3. mining frequent patterns
3. mining frequent patterns
 
3.3 hierarchical methods
3.3 hierarchical methods3.3 hierarchical methods
3.3 hierarchical methods
 
Pca
PcaPca
Pca
 
Presentation on unsupervised learning
Presentation on unsupervised learning Presentation on unsupervised learning
Presentation on unsupervised learning
 
K means Clustering Algorithm
K means Clustering AlgorithmK means Clustering Algorithm
K means Clustering Algorithm
 
Data cleaning-outlier-detection
Data cleaning-outlier-detectionData cleaning-outlier-detection
Data cleaning-outlier-detection
 
Optimization/Gradient Descent
Optimization/Gradient DescentOptimization/Gradient Descent
Optimization/Gradient Descent
 
3.2 partitioning methods
3.2 partitioning methods3.2 partitioning methods
3.2 partitioning methods
 
Dbscan algorithom
Dbscan algorithomDbscan algorithom
Dbscan algorithom
 
Clusters techniques
Clusters techniquesClusters techniques
Clusters techniques
 
07 dimensionality reduction
07 dimensionality reduction07 dimensionality reduction
07 dimensionality reduction
 
Image segmentation
Image segmentationImage segmentation
Image segmentation
 
Random forest algorithm
Random forest algorithmRandom forest algorithm
Random forest algorithm
 
Naive Bayes
Naive BayesNaive Bayes
Naive Bayes
 
CIFAR-10
CIFAR-10CIFAR-10
CIFAR-10
 
Types of clustering and different types of clustering algorithms
Types of clustering and different types of clustering algorithmsTypes of clustering and different types of clustering algorithms
Types of clustering and different types of clustering algorithms
 
Unsupervised learning (clustering)
Unsupervised learning (clustering)Unsupervised learning (clustering)
Unsupervised learning (clustering)
 
04 Classification in Data Mining
04 Classification in Data Mining04 Classification in Data Mining
04 Classification in Data Mining
 
Hierarchical clustering.pptx
Hierarchical clustering.pptxHierarchical clustering.pptx
Hierarchical clustering.pptx
 

Viewers also liked

Linear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its GeneralizationLinear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its Generalization일상 온
 
face recognition system
face recognition systemface recognition system
face recognition systemAnil Kumar
 
LDA Beginner's Tutorial
LDA Beginner's TutorialLDA Beginner's Tutorial
LDA Beginner's TutorialWayne Lee
 
Some contributions to PCA for time series
Some contributions to PCA for time seriesSome contributions to PCA for time series
Some contributions to PCA for time seriesIsabel Magalhães
 
Latent Dirichlet Allocation
Latent Dirichlet AllocationLatent Dirichlet Allocation
Latent Dirichlet AllocationMarco Righini
 
Topic Models, LDA and all that
Topic Models, LDA and all thatTopic Models, LDA and all that
Topic Models, LDA and all thatZhibo Xiao
 
LDA presentation
LDA presentationLDA presentation
LDA presentationMohit Gupta
 
Text Mining using LDA with Context
Text Mining using LDA with ContextText Mining using LDA with Context
Text Mining using LDA with ContextSteffen Staab
 
PCA Based Face Recognition System
PCA Based Face Recognition SystemPCA Based Face Recognition System
PCA Based Face Recognition SystemMd. Atiqur Rahman
 
Face recognition technology - BEST PPT
Face recognition technology - BEST PPTFace recognition technology - BEST PPT
Face recognition technology - BEST PPTSiddharth Modi
 
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vecword2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec👋 Christopher Moody
 
2012 mdsp pr09 pca lda
2012 mdsp pr09 pca lda2012 mdsp pr09 pca lda
2012 mdsp pr09 pca ldanozomuhamada
 
Big Data Analysis: The curse of dimensionality in official statistics
Big Data Analysis: The curse of dimensionality in official statisticsBig Data Analysis: The curse of dimensionality in official statistics
Big Data Analysis: The curse of dimensionality in official statisticsDario Buono
 

Viewers also liked (20)

PCA vs LDA
PCA vs LDAPCA vs LDA
PCA vs LDA
 
Face recognition using LDA
Face recognition using LDAFace recognition using LDA
Face recognition using LDA
 
Linear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its GeneralizationLinear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its Generalization
 
face recognition system
face recognition systemface recognition system
face recognition system
 
Lda
LdaLda
Lda
 
LDA Beginner's Tutorial
LDA Beginner's TutorialLDA Beginner's Tutorial
LDA Beginner's Tutorial
 
Some contributions to PCA for time series
Some contributions to PCA for time seriesSome contributions to PCA for time series
Some contributions to PCA for time series
 
LDA
LDALDA
LDA
 
Latent Dirichlet Allocation
Latent Dirichlet AllocationLatent Dirichlet Allocation
Latent Dirichlet Allocation
 
Topic Models, LDA and all that
Topic Models, LDA and all thatTopic Models, LDA and all that
Topic Models, LDA and all that
 
LDA presentation
LDA presentationLDA presentation
LDA presentation
 
Text Mining using LDA with Context
Text Mining using LDA with ContextText Mining using LDA with Context
Text Mining using LDA with Context
 
Principal component analysis
Principal component analysisPrincipal component analysis
Principal component analysis
 
PCA Based Face Recognition System
PCA Based Face Recognition SystemPCA Based Face Recognition System
PCA Based Face Recognition System
 
Principal Component Analysis
Principal Component AnalysisPrincipal Component Analysis
Principal Component Analysis
 
07. PCA
07. PCA07. PCA
07. PCA
 
Face recognition technology - BEST PPT
Face recognition technology - BEST PPTFace recognition technology - BEST PPT
Face recognition technology - BEST PPT
 
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vecword2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec
 
2012 mdsp pr09 pca lda
2012 mdsp pr09 pca lda2012 mdsp pr09 pca lda
2012 mdsp pr09 pca lda
 
Big Data Analysis: The curse of dimensionality in official statistics
Big Data Analysis: The curse of dimensionality in official statisticsBig Data Analysis: The curse of dimensionality in official statistics
Big Data Analysis: The curse of dimensionality in official statistics
 

Similar to Understandig PCA and LDA

Fr pca lda
Fr pca ldaFr pca lda
Fr pca ldaultraraj
 
DimensionalityReduction.pptx
DimensionalityReduction.pptxDimensionalityReduction.pptx
DimensionalityReduction.pptx36rajneekant
 
Dimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptxDimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptxSivam Chinna
 
pcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptxpcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptxABINASHPADHY6
 
introduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptxintroduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptxPratik Gohel
 
Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Mostafa G. M. Mostafa
 
Matrix Factorization In Recommender Systems
Matrix Factorization In Recommender SystemsMatrix Factorization In Recommender Systems
Matrix Factorization In Recommender SystemsYONG ZHENG
 
5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdf5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdfRahul926331
 
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image DescriptorsPCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image Descriptorswolf
 
Singular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptxSingular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptxrajalakshmi5921
 
EDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptxEDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptxrajalakshmi5921
 
The following ppt is about principal component analysis
The following ppt is about principal component analysisThe following ppt is about principal component analysis
The following ppt is about principal component analysisSushmit8
 
Computational Giants_nhom.pptx
Computational Giants_nhom.pptxComputational Giants_nhom.pptx
Computational Giants_nhom.pptxThAnhonc
 
Face Recognition Using Neural Networks
Face Recognition Using Neural NetworksFace Recognition Using Neural Networks
Face Recognition Using Neural NetworksCSCJournals
 
Aaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reductionAaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reductionAminaRepo
 

Similar to Understandig PCA and LDA (20)

Fr pca lda
Fr pca ldaFr pca lda
Fr pca lda
 
DimensionalityReduction.pptx
DimensionalityReduction.pptxDimensionalityReduction.pptx
DimensionalityReduction.pptx
 
Dimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptxDimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptx
 
pcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptxpcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptx
 
introduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptxintroduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptx
 
Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)
 
Matrix Factorization In Recommender Systems
Matrix Factorization In Recommender SystemsMatrix Factorization In Recommender Systems
Matrix Factorization In Recommender Systems
 
5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdf5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdf
 
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image DescriptorsPCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
 
Singular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptxSingular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptx
 
EDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptxEDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptx
 
pca.ppt
pca.pptpca.ppt
pca.ppt
 
The following ppt is about principal component analysis
The following ppt is about principal component analysisThe following ppt is about principal component analysis
The following ppt is about principal component analysis
 
PCA Final.pptx
PCA Final.pptxPCA Final.pptx
PCA Final.pptx
 
Fa18_P2.pptx
Fa18_P2.pptxFa18_P2.pptx
Fa18_P2.pptx
 
Machine learning meetup
Machine learning meetupMachine learning meetup
Machine learning meetup
 
Computational Giants_nhom.pptx
Computational Giants_nhom.pptxComputational Giants_nhom.pptx
Computational Giants_nhom.pptx
 
Face Recognition Using Neural Networks
Face Recognition Using Neural NetworksFace Recognition Using Neural Networks
Face Recognition Using Neural Networks
 
Aaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reductionAaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reduction
 
CSC446: Pattern Recognition (LN6)
CSC446: Pattern Recognition (LN6)CSC446: Pattern Recognition (LN6)
CSC446: Pattern Recognition (LN6)
 

More from Dr. Syed Hassan Amin

Greenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparisonGreenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparisonDr. Syed Hassan Amin
 
Multitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq RecognitionMultitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq RecognitionDr. Syed Hassan Amin
 
Thin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better CodeThin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better CodeDr. Syed Hassan Amin
 
Improving Code Quality Through Effective Review Process
Improving Code Quality Through Effective  Review ProcessImproving Code Quality Through Effective  Review Process
Improving Code Quality Through Effective Review ProcessDr. Syed Hassan Amin
 
Software Project Management Tips and Tricks
Software Project Management Tips and TricksSoftware Project Management Tips and Tricks
Software Project Management Tips and TricksDr. Syed Hassan Amin
 
Improving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design PrinciplesImproving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design PrinciplesDr. Syed Hassan Amin
 
Learning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve JobsLearning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve JobsDr. Syed Hassan Amin
 
Understanding and Managing Technical Debt
Understanding and Managing Technical DebtUnderstanding and Managing Technical Debt
Understanding and Managing Technical DebtDr. Syed Hassan Amin
 
An OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq FontAn OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq FontDr. Syed Hassan Amin
 

More from Dr. Syed Hassan Amin (12)

Greenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparisonGreenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparison
 
Introduction To Docker
Introduction To  DockerIntroduction To  Docker
Introduction To Docker
 
Laravel Unit Testing
Laravel Unit TestingLaravel Unit Testing
Laravel Unit Testing
 
Multitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq RecognitionMultitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq Recognition
 
Agile Scrum Methodology
Agile Scrum MethodologyAgile Scrum Methodology
Agile Scrum Methodology
 
Thin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better CodeThin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better Code
 
Improving Code Quality Through Effective Review Process
Improving Code Quality Through Effective  Review ProcessImproving Code Quality Through Effective  Review Process
Improving Code Quality Through Effective Review Process
 
Software Project Management Tips and Tricks
Software Project Management Tips and TricksSoftware Project Management Tips and Tricks
Software Project Management Tips and Tricks
 
Improving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design PrinciplesImproving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design Principles
 
Learning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve JobsLearning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve Jobs
 
Understanding and Managing Technical Debt
Understanding and Managing Technical DebtUnderstanding and Managing Technical Debt
Understanding and Managing Technical Debt
 
An OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq FontAn OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq Font
 

Recently uploaded

High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...soniya singh
 
20240419 - Measurecamp Amsterdam - SAM.pdf
20240419 - Measurecamp Amsterdam - SAM.pdf20240419 - Measurecamp Amsterdam - SAM.pdf
20240419 - Measurecamp Amsterdam - SAM.pdfHuman37
 
Industrialised data - the key to AI success.pdf
Industrialised data - the key to AI success.pdfIndustrialised data - the key to AI success.pdf
Industrialised data - the key to AI success.pdfLars Albertsson
 
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝soniya singh
 
RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998YohFuh
 
From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...Florian Roscheck
 
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...Suhani Kapoor
 
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...Sapana Sha
 
100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptxAnupama Kate
 
Data Science Project: Advancements in Fetal Health Classification
Data Science Project: Advancements in Fetal Health ClassificationData Science Project: Advancements in Fetal Health Classification
Data Science Project: Advancements in Fetal Health ClassificationBoston Institute of Analytics
 
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptdokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptSonatrach
 
Spark3's new memory model/management
Spark3's new memory model/managementSpark3's new memory model/management
Spark3's new memory model/managementakshesh doshi
 
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改atducpo
 
04242024_CCC TUG_Joins and Relationships
04242024_CCC TUG_Joins and Relationships04242024_CCC TUG_Joins and Relationships
04242024_CCC TUG_Joins and Relationshipsccctableauusergroup
 
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...Suhani Kapoor
 
Schema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfSchema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfLars Albertsson
 

Recently uploaded (20)

High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
 
20240419 - Measurecamp Amsterdam - SAM.pdf
20240419 - Measurecamp Amsterdam - SAM.pdf20240419 - Measurecamp Amsterdam - SAM.pdf
20240419 - Measurecamp Amsterdam - SAM.pdf
 
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
꧁❤ Aerocity Call Girls Service Aerocity Delhi ❤꧂ 9999965857 ☎️ Hard And Sexy ...
 
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in Kishangarh
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in  KishangarhDelhi 99530 vip 56974 Genuine Escort Service Call Girls in  Kishangarh
Delhi 99530 vip 56974 Genuine Escort Service Call Girls in Kishangarh
 
Industrialised data - the key to AI success.pdf
Industrialised data - the key to AI success.pdfIndustrialised data - the key to AI success.pdf
Industrialised data - the key to AI success.pdf
 
E-Commerce Order PredictionShraddha Kamble.pptx
E-Commerce Order PredictionShraddha Kamble.pptxE-Commerce Order PredictionShraddha Kamble.pptx
E-Commerce Order PredictionShraddha Kamble.pptx
 
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝
Call Girls in Defence Colony Delhi 💯Call Us 🔝8264348440🔝
 
RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998
 
From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...
 
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
 
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
 
100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx
 
Data Science Project: Advancements in Fetal Health Classification
Data Science Project: Advancements in Fetal Health ClassificationData Science Project: Advancements in Fetal Health Classification
Data Science Project: Advancements in Fetal Health Classification
 
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptdokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
 
Spark3's new memory model/management
Spark3's new memory model/managementSpark3's new memory model/management
Spark3's new memory model/management
 
Russian Call Girls Dwarka Sector 15 💓 Delhi 9999965857 @Sabina Modi VVIP MODE...
Russian Call Girls Dwarka Sector 15 💓 Delhi 9999965857 @Sabina Modi VVIP MODE...Russian Call Girls Dwarka Sector 15 💓 Delhi 9999965857 @Sabina Modi VVIP MODE...
Russian Call Girls Dwarka Sector 15 💓 Delhi 9999965857 @Sabina Modi VVIP MODE...
 
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
 
04242024_CCC TUG_Joins and Relationships
04242024_CCC TUG_Joins and Relationships04242024_CCC TUG_Joins and Relationships
04242024_CCC TUG_Joins and Relationships
 
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
 
Schema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfSchema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdf
 

Understandig PCA and LDA

  • 1. CS 312 Character and Pattern Recognition – Spring 2009 Dimensionality Reduction Using PCA/LDA Chapter 3 (Duda et al.) – Section 3.8 Thanks to Prof. Bebis Case Studies: Face Recognition Using Dimensionality Reduction M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive Neuroscience, 3(1), pp. 71-86, 1991. D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image Retrieval", IEEE Transactions on Pattern Analysis and Machine Intelligence, 18(8), pp. 831-836, 1996. A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp. 228-233, 2001.
  • 2. 2 Dimensionality Reduction • One approach to deal with high dimensional data is by reducing their dimensionality. • Project high dimensional data onto a lower dimensional sub- space using linear or non-linear transformations.
  • 3. 3 Dimensionality Reduction • Linear transformations are simple to compute and tractable. • Classical –linear- approaches: – Principal Component Analysis (PCA) – Fisher Discriminant Analysis (FDA) ( )t i i iY U X b u a= = k x 1 k x d d x 1 (k<<d)k x 1 k x d d x 1 (k<<d)
  • 4. 4 Principal Component Analysis (PCA) • Each dimensionality reduction technique finds an appropriate transformation by satisfying certain criteria (e.g., information loss, data discrimination, etc.) • The goal of PCA is to reduce the dimensionality of the data while retaining as much as possible of the variation present in the dataset.
  • 5. 5 Principal Component Analysis (PCA) 1 1 2 2 1 2 ˆ ... where , ,..., isa basein the -dimensionalsub-space(K<N) K K K x b u b u b u u u u K = + + + ˆx x= 1 1 2 2 1 2 ... where , ,..., isa basein theoriginal N-dimensionalspace N N n x a v a v a v v v v = + + + • Find a basis in a low dimensional sub-space: − Approximate vectors by projecting them in a low dimensional sub-space: (1) Original space representation: (2) Lower-dimensional sub-space representation: • Note: if K=N, then
  • 6. 6 Principal Component Analysis (PCA) • Example (K=N):
  • 7. 7 Principal Component Analysis (PCA) • Information loss − Dimensionality reduction implies information loss !! − PCA preserves as much information as possible: • What is the “best” lower dimensional sub-space? The “best” low-dimensional space is centered at the sample mean and has directions determined by the “best” eigenvectors of the covariance matrix of the data x. − By “best” eigenvectors we mean those corresponding to the largest eigenvalues ( i.e., “principal components”). − Since the covariance matrix is real and symmetric, these eigenvectors are orthogonal and form a set of basis vectors. (see pp. 114-117 in textbook for a proof) ˆmin || || (reconstruction error)x x−
  • 8. 8 Principal Component Analysis (PCA) • Methodology − Suppose x1, x2, ..., xM are N x 1 vectors
  • 9. 9 Principal Component Analysis (PCA) • Methodology – cont. ( )T i ib u x x= −
  • 10. 10 Principal Component Analysis (PCA) • Eigenvalue spectrum λi K λN
  • 11. 11 Principal Component Analysis (PCA) • Linear transformation implied by PCA − The linear transformation RN → RK that performs the dimensionality reduction is:
  • 12. 12 Principal Component Analysis (PCA) • Geometric interpretation − PCA projects the data along the directions where the data varies the most. − These directions are determined by the eigenvectors of the covariance matrix corresponding to the largest eigenvalues. − The magnitude of the eigenvalues corresponds to the variance of the data along the eigenvector directions.
  • 13. 13 Principal Component Analysis (PCA) • How many principal components to keep? − To choose K, you can use the following criterion:
  • 14. 14 Principal Component Analysis (PCA) • What is the error due to dimensionality reduction? • It can be shown that the average error due to dimensionality reduction is equal to: e
  • 15. 15 Principal Component Analysis (PCA) • Standardization − The principal components are dependent on the units used to measure the original variables as well as on the range of values they assume. − We should always standardize the data prior to using PCA. − A common standardization method is to transform all the data to have zero mean and unit standard deviation:
  • 16. 16 Principal Component Analysis (PCA) • Case Study: Eigenfaces for Face Detection/Recognition − M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive Neuroscience, vol. 3, no. 1, pp. 71-86, 1991. • Face Recognition − The simplest approach is to think of it as a template matching problem − Problems arise when performing recognition in a high- dimensional space. − Significant improvements can be achieved by first mapping the data into a lower dimensionality space. − How to find this lower-dimensional space?
  • 17. 17 Principal Component Analysis (PCA) • Main idea behind eigenfaces ˆΦ −Ψ −Ψ average face
  • 18. 18 Principal Component Analysis (PCA) • Computation of the eigenfaces
  • 19. 19 Principal Component Analysis (PCA) • Computation of the eigenfaces – cont.
  • 20. 20 Principal Component Analysis (PCA) • Computation of the eigenfaces – cont. ui
  • 21. 21 Principal Component Analysis (PCA) • Computation of the eigenfaces – cont.
  • 22. 22 Principal Component Analysis (PCA) • Representing faces onto this basis ˆ iΦ
  • 23. 23 Principal Component Analysis (PCA) • Eigenvalue spectrum : Defined as set of Eigenvalues. • Spectrum is shown as Eigenvalues plotted in the graph shown below: λi K M λN
  • 24. 24 Principal Component Analysis (PCA) • Representing faces onto this basis – cont.
  • 25. 25 Principal Component Analysis (PCA) • Face Recognition Using Eigenfaces
  • 26. 26 Principal Component Analysis (PCA) • Face Recognition Using Eigenfaces – cont. − The distance er is called distance within the face space (difs) − Comment: we can use the common Euclidean distance to compute er, however, it has been reported that the Mahalanobis distance performs better:
  • 27. 27 Principal Component Analysis (PCA) • Face Detection Using Eigenfaces
  • 28. 28 Principal Component Analysis (PCA) • Face Detection Using Eigenfaces – cont.
  • 29. 29 Principal Component Analysis (PCA) • Reconstruction of faces and non-faces
  • 30. 30 Principal Component Analysis (PCA) • Applications − Face detection, tracking, and recognition dffs
  • 31. 31 Principal Component Analysis (PCA) • Problems in face recognition: − Background (de-emphasize the outside of the face – e.g., by multiplying the input image by a 2D Gaussian window centered on the face) − Lighting conditions (performance degrades with light changes) − Scale (performance decreases quickly with changes to head size) − multi-scale eigenspaces − scale input image to multiple sizes − Orientation (performance decreases but not as fast as with scale changes) − plane rotations can be handled − out-of-plane rotations are more difficult to handle
  • 32. 32 Principal Component Analysis (PCA) • Dataset − 16 subjects − 3 orientations, 3 sizes − 3 lighting conditions, 6 resolutions (512x512 ... 16x16) − Total number of images: 2,592
  • 33. 33 Principal Component Analysis (PCA) • Experiment 1 − Used various sets of 16 images for training − One image/person, taken under the same conditions − Classify the rest images as one of the 16 individuals − 7 eigenfaces were used − No rejections (i.e., no threshold on difs) − Performed a large number of experiments and averaged the results: − 96% correct averaged over light variation − 85% correct averaged over orientation variation − 64% correct averaged over size variation
  • 34. 34 Principal Component Analysis (PCA) • Experiment 2 − Considered rejections (i.e., by thresholding difs) − There is a tradeoff between correct recognition and rejections. − Adjusting the threshold to achieve 100% recognition accuracy resulted in: − 19% rejections while varying lighting − 39% rejections while varying orientation − 60% rejections while varying size • Experiment 3 − Reconstruction using partial information
  • 35. 35 Principal Component Analysis (PCA) • PCA and classification − PCA is not always an optimal dimensionality-reduction procedure for classification purposes. • Multiple classes and PCA − Suppose there are C classes in the training data. − PCA is based on the sample covariance which characterizes the scatter of the entire data set, irrespective of class-membership. − The projection axes chosen by PCA might not provide good discrimination power.
  • 36. 36 Principal Component Analysis (PCA) • PCA and classification (cont’d)
  • 37. 37 Linear Discriminant Analysis (LDA) • What is the goal of LDA? − Perform dimensionality reduction “while preserving as much of the class discriminatory information as possible”. − Seeks to find directions along which the classes are best separated. − Takes into consideration the scatter within-classes but also the scatter between-classes. − More capable of distinguishing image variation due to identity from variation due to other sources such as illumination and expression.
  • 39. 39 Linear Discriminant Analysis (LDA) • Notation (Sb has at most rank C-1) (each sub-matrix has rank 1 or less, i.e., outer product of two vectors) 1 1 ( )( ) iMC T w j i j i i j S x x = = = − −∑∑ μ μ
  • 40. 40 Linear Discriminant Analysis (LDA) • Methodology − LDA computes a transformation that maximizes the between-class scatter while minimizing the within-class scatter: | | | | max max | | | | T b b T w w U S U S U S U S = % % T U=y x products of eigenvalues ! projection matrix ,b wS S% % : scatter matrices of the projected data y
  • 41. 41 Linear Discriminant Analysis (LDA) • Linear transformation implied by LDA − The linear transformation is given by a matrix U whose columns are the eigenvectors of the above problem (i.e., called Fisherfaces). − The LDA solution is given by the eigenvectors of the generalized eigenvector problem: − Important: Since Sb has at most rank C-1, the max number of eigenvectors with non-zero eigenvalues is C-1 (i.e., max dimensionality of sub-space is C-1)
  • 42. 42 Linear Discriminant Analysis (LDA) • Does Sw -1 always exist? − If Sw is non-singular, we can obtain a conventional eigenvalue problem by writing: − In practice, Sw is often singular since the data are image vectors with large dimensionality while the size of the data set is much smaller (M << N )
  • 43. 43 Linear Discriminant Analysis (LDA) • Does Sw -1 always exist? – cont. − To alleviate this problem, we can use PCA first: 1) PCA is first applied to the data set to reduce its dimensionality. 2) LDA is then applied to find the most discriminative directions:
  • 44. 44 Linear Discriminant Analysis (LDA) • Case Study: Using Discriminant Eigenfeatures for Image Retrieval − D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image Retrieval", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 18, no. 8, pp. 831-836, 1996. • Content-based image retrieval − The application being studied here is query-by-example image retrieval. − The paper deals with the problem of selecting a good set of image features for content-based image retrieval.
  • 45. 45 Linear Discriminant Analysis (LDA) • Assumptions − "Well-framed" images are required as input for training and query- by-example test probes. − Only a small variation in the size, position, and orientation of the objects in the images is allowed.
  • 46. 46 Linear Discriminant Analysis (LDA) • Some terminology − Most Expressive Features (MEF): the features (projections) obtained using PCA. − Most Discriminating Features (MDF): the features (projections) obtained using LDA. • Numerical problems − When computing the eigenvalues/eigenvectors of Sw -1 SBuk = λkuk numerically, the computations can be unstable since Sw -1 SB is not always symmetric. − See paper for a way to find the eigenvalues/eigenvectors in a stable way.
  • 47. 47 Linear Discriminant Analysis (LDA) • Factors unrelated to classification − MEF vectors show the tendency of PCA to capture major variations in the training set such as lighting direction. − MDF vectors discount those factors unrelated to classification.
  • 48. 48 Linear Discriminant Analysis (LDA) • Clustering effect 1) Generate the set of MEFs/MDFs for each image in the training set. 2) Given an query image, compute its MEFs/MDFs using the same procedure. 3) Find the k closest neighbors for retrieval (e.g., using Euclidean distance). • Methodology
  • 49. 49 Linear Discriminant Analysis (LDA) • Experiments and results − Face images − A set of face images was used with 2 expressions, 3 lighting conditions. − Testing was performed using a disjoint set of images: • One image, randomly chosen, from each individual.
  • 51. 51 Linear Discriminant Analysis (LDA) − Examples of correct search probes
  • 52. 52 Linear Discriminant Analysis (LDA) − Example of a failed search probe
  • 53. 53 Linear Discriminant Analysis (LDA) • Case Study: PCA versus LDA − A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp. 228- 233, 2001. • Is LDA always better than PCA? − There has been a tendency in the computer vision community to prefer LDA over PCA. − This is mainly because LDA deals directly with discrimination between classes while PCA does not pay attention to the underlying class structure. − Main results of this study: (1) When the training set is small, PCA can outperform LDA. (2) When the number of samples is large and representative for each class, LDA outperforms PCA.
  • 54. 54 Linear Discriminant Analysis (LDA) • Is LDA always better than PCA? – cont.
  • 55. 55 Linear Discriminant Analysis (LDA) • Is LDA always better than PCA? – cont. LDA is not always better when training set is small
  • 56. 56 Linear Discriminant Analysis (LDA) • Is LDA always better than PCA? – cont. LDA outperforms PCA when training set is large

Editor's Notes

  1. Good afternoon and thank you everyone for coming. My talk today will describe the research I performed at the IRIS at USC, the object of this work being to build a computational framework that addresses the problem of motion analysis and interpretation.