SlideShare une entreprise Scribd logo
1  sur  57
CS 312 Character and
Pattern Recognition – Spring 2009
Dimensionality Reduction Using PCA/LDA
Chapter 3 (Duda et al.) – Section 3.8
Thanks to Prof. Bebis
Case Studies:
Face Recognition Using Dimensionality Reduction
M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive Neuroscience, 3(1),
pp. 71-86, 1991.
D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image Retrieval", IEEE
Transactions on Pattern Analysis and Machine Intelligence, 18(8), pp. 831-836, 1996.
A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on Pattern Analysis and
Machine Intelligence, vol. 23, no. 2, pp. 228-233, 2001.
2
Dimensionality Reduction
• One approach to deal with high dimensional data is by reducing
their dimensionality.
• Project high dimensional data onto a lower dimensional sub-
space using linear or non-linear transformations.
3
Dimensionality Reduction
• Linear transformations are simple to compute and tractable.
• Classical –linear- approaches:
– Principal Component Analysis (PCA)
– Fisher Discriminant Analysis (FDA)
( )t
i i iY U X b u a= =
k x 1 k x d d x 1 (k<<d)k x 1 k x d d x 1 (k<<d)
4
Principal Component Analysis (PCA)
• Each dimensionality reduction technique finds an
appropriate transformation by satisfying certain criteria
(e.g., information loss, data discrimination, etc.)
• The goal of PCA is to reduce the dimensionality of the data
while retaining as much as possible of the variation
present in the dataset.
5
Principal Component Analysis (PCA)
1 1 2 2
1 2
ˆ ...
where , ,..., isa basein the -dimensionalsub-space(K<N)
K K
K
x b u b u b u
u u u K
= + + +
ˆx x=
1 1 2 2
1 2
...
where , ,..., isa basein theoriginal N-dimensionalspace
N N
n
x a v a v a v
v v v
= + + +
• Find a basis in a low dimensional sub-space:
− Approximate vectors by projecting them in a low dimensional
sub-space:
(1) Original space representation:
(2) Lower-dimensional sub-space representation:
• Note: if K=N, then
6
Principal Component Analysis (PCA)
• Example (K=N):
7
Principal Component Analysis (PCA)
• Information loss
− Dimensionality reduction implies information loss !!
− PCA preserves as much information as possible:
• What is the “best” lower dimensional sub-space?
The “best” low-dimensional space is centered at the sample mean
and has directions determined by the “best” eigenvectors of the
covariance matrix of the data x.
− By “best” eigenvectors we mean those corresponding to the largest
eigenvalues ( i.e., “principal components”).
− Since the covariance matrix is real and symmetric, these
eigenvectors are orthogonal and form a set of basis vectors.
(see pp. 114-117 in textbook for a proof)
ˆmin || || (reconstruction error)x x−
8
Principal Component Analysis (PCA)
• Methodology
− Suppose x1, x2, ..., xM are N x 1 vectors
9
Principal Component Analysis (PCA)
• Methodology – cont.
( )T
i ib u x x= −
10
Principal Component Analysis (PCA)
• Eigenvalue spectrum
λi
K
λN
11
Principal Component Analysis (PCA)
• Linear transformation implied by PCA
− The linear transformation RN
→ RK
that performs the dimensionality
reduction is:
12
Principal Component Analysis (PCA)
• Geometric interpretation
− PCA projects the data along the directions where the data varies the
most.
− These directions are determined by the eigenvectors of the
covariance matrix corresponding to the largest eigenvalues.
− The magnitude of the eigenvalues corresponds to the variance of
the data along the eigenvector directions.
13
Principal Component Analysis (PCA)
• How many principal components to keep?
− To choose K, you can use the following criterion:
14
Principal Component Analysis (PCA)
• What is the error due to dimensionality reduction?
• It can be shown that the average error due to
dimensionality reduction is equal to:
e
15
Principal Component Analysis (PCA)
• Standardization
− The principal components are dependent on the units used to
measure the original variables as well as on the range of values
they assume.
− We should always standardize the data prior to using PCA.
− A common standardization method is to transform all the data to
have zero mean and unit standard deviation:
16
Principal Component Analysis (PCA)
• Case Study: Eigenfaces for Face Detection/Recognition
− M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of
Cognitive Neuroscience, vol. 3, no. 1, pp. 71-86, 1991.
• Face Recognition
− The simplest approach is to think of it as a template matching
problem
− Problems arise when performing recognition in a high-
dimensional space.
− Significant improvements can be achieved by first mapping
the data into a lower dimensionality space.
− How to find this lower-dimensional space?
17
Principal Component Analysis (PCA)
• Main idea behind eigenfaces
ˆΦ −Ψ
−Ψ
average face
18
Principal Component Analysis (PCA)
• Computation of the eigenfaces
19
Principal Component Analysis (PCA)
• Computation of the eigenfaces – cont.
20
Principal Component Analysis (PCA)
• Computation of the eigenfaces – cont.
ui
21
Principal Component Analysis (PCA)
• Computation of the eigenfaces – cont.
22
Principal Component Analysis (PCA)
• Representing faces onto this basis
ˆ
iΦ
23
Principal Component Analysis (PCA)
• Eigenvalue spectrum : Defined as set of Eigenvalues.
• Spectrum is shown as Eigenvalues plotted in the graph
shown below:
λi
K
M
λN
24
Principal Component Analysis (PCA)
• Representing faces onto this basis – cont.
25
Principal Component Analysis (PCA)
• Face Recognition Using Eigenfaces
26
Principal Component Analysis (PCA)
• Face Recognition Using Eigenfaces – cont.
− The distance er is called distance within the face space (difs)
− Comment: we can use the common Euclidean distance to compute
er, however, it has been reported that the Mahalanobis distance
performs better:
27
Principal Component Analysis (PCA)
• Face Detection Using Eigenfaces
28
Principal Component Analysis (PCA)
• Face Detection Using Eigenfaces – cont.
29
Principal Component Analysis (PCA)
• Reconstruction of
faces and non-faces
30
Principal Component Analysis (PCA)
• Applications
− Face detection, tracking, and recognition
dffs
31
Principal Component Analysis (PCA)
• Problems in face recognition:
− Background (de-emphasize the outside of the face – e.g., by
multiplying the input image by a 2D Gaussian window centered on
the face)
− Lighting conditions (performance degrades with light changes)
− Scale (performance decreases quickly with changes to head size)
− multi-scale eigenspaces
− scale input image to multiple sizes
− Orientation (performance decreases but not as fast as with scale
changes)
− plane rotations can be handled
− out-of-plane rotations are more difficult to handle
32
Principal Component Analysis (PCA)
• Dataset
− 16 subjects
− 3 orientations, 3 sizes
− 3 lighting conditions, 6 resolutions (512x512 ... 16x16)
− Total number of images: 2,592
33
Principal Component Analysis (PCA)
• Experiment 1
− Used various sets of 16 images for training
− One image/person, taken under the same conditions
− Classify the rest images as one of the 16 individuals
− 7 eigenfaces were used
− No rejections (i.e., no threshold on difs)
− Performed a large number of experiments and averaged the results:
− 96% correct averaged over light variation
− 85% correct averaged over orientation variation
− 64% correct averaged over size variation
34
Principal Component Analysis (PCA)
• Experiment 2
− Considered rejections (i.e., by thresholding difs)
− There is a tradeoff between correct recognition and rejections.
− Adjusting the threshold to achieve 100% recognition accuracy
resulted in:
− 19% rejections while varying lighting
− 39% rejections while varying orientation
− 60% rejections while varying size
• Experiment 3
− Reconstruction using
partial information
35
Principal Component Analysis (PCA)
• PCA and classification
− PCA is not always an optimal dimensionality-reduction procedure
for classification purposes.
• Multiple classes and PCA
− Suppose there are C classes in the training data.
− PCA is based on the sample covariance which characterizes the
scatter of the entire data set, irrespective of class-membership.
− The projection axes chosen by PCA might not provide good
discrimination power.
36
Principal Component Analysis (PCA)
• PCA and classification (cont’d)
37
Linear Discriminant Analysis (LDA)
• What is the goal of LDA?
− Perform dimensionality reduction “while preserving as much of
the class discriminatory information as possible”.
− Seeks to find directions along which the classes are best separated.
− Takes into consideration the scatter within-classes but also the
scatter between-classes.
− More capable of distinguishing image variation due to identity from
variation due to other sources such as illumination and expression.
38
Linear Discriminant Analysis (LDA)
39
Linear Discriminant Analysis (LDA)
• Notation
(Sb has at most rank C-1)
(each sub-matrix has
rank 1 or less, i.e., outer
product of two vectors)
1 1
( )( )
iMC
T
w j i j i
i j
S x x
= =
= − −∑∑ μ μ
40
Linear Discriminant Analysis (LDA)
• Methodology
− LDA computes a transformation that maximizes the between-class
scatter while minimizing the within-class scatter:
| | | |
max max
| | | |
T
b b
T
w w
U S U S
U S U S
=
%
%
T
U=y x
products of eigenvalues !
projection matrix
,b wS S% % : scatter matrices of the projected data y
41
Linear Discriminant Analysis (LDA)
• Linear transformation implied by LDA
− The linear transformation is given by a matrix U whose columns are
the eigenvectors of the above problem (i.e., called Fisherfaces).
− The LDA solution is given by the eigenvectors of the generalized
eigenvector problem:
− Important: Since Sb has at most rank C-1, the max number of
eigenvectors with non-zero eigenvalues is C-1 (i.e., max
dimensionality of sub-space is C-1)
42
Linear Discriminant Analysis (LDA)
• Does Sw
-1
always exist?
− If Sw is non-singular, we can obtain a conventional eigenvalue
problem by writing:
− In practice, Sw is often singular since the data are image vectors
with large dimensionality while the size of the data set is much
smaller (M << N )
43
Linear Discriminant Analysis (LDA)
• Does Sw
-1
always exist? – cont.
− To alleviate this problem, we can use PCA first:
1) PCA is first applied to the data set to reduce its dimensionality.
2) LDA is then applied to find the most discriminative directions:
44
Linear Discriminant Analysis (LDA)
• Case Study: Using Discriminant Eigenfeatures for Image
Retrieval
− D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image
Retrieval", IEEE Transactions on Pattern Analysis and Machine
Intelligence, vol. 18, no. 8, pp. 831-836, 1996.
• Content-based image retrieval
− The application being studied here is query-by-example image
retrieval.
− The paper deals with the problem of selecting a good set of image
features for content-based image retrieval.
45
Linear Discriminant Analysis (LDA)
• Assumptions
− "Well-framed" images are required as input for training and query-
by-example test probes.
− Only a small variation in the size, position, and orientation of the
objects in the images is allowed.
46
Linear Discriminant Analysis (LDA)
• Some terminology
− Most Expressive Features (MEF): the features (projections)
obtained using PCA.
− Most Discriminating Features (MDF): the features (projections)
obtained using LDA.
• Numerical problems
− When computing the eigenvalues/eigenvectors of Sw
-1
SBuk = λkuk
numerically, the computations can be unstable since Sw
-1
SB is not
always symmetric.
− See paper for a way to find the eigenvalues/eigenvectors in a stable
way.
47
Linear Discriminant Analysis (LDA)
• Factors unrelated to classification
− MEF vectors show the tendency of PCA to capture major variations
in the training set such as lighting direction.
− MDF vectors discount those factors unrelated to classification.
48
Linear Discriminant Analysis (LDA)
• Clustering effect
1) Generate the set of MEFs/MDFs for each image in the training set.
2) Given an query image, compute its MEFs/MDFs using the same
procedure.
3) Find the k closest neighbors for retrieval (e.g., using Euclidean
distance).
• Methodology
49
Linear Discriminant Analysis (LDA)
• Experiments and results
− Face images
− A set of face images was used with 2 expressions, 3 lighting conditions.
− Testing was performed using a disjoint set of images:
• One image, randomly chosen, from each individual.
50
Linear Discriminant Analysis (LDA)
51
Linear Discriminant Analysis (LDA)
− Examples of correct search probes
52
Linear Discriminant Analysis (LDA)
− Example of a failed search probe
53
Linear Discriminant Analysis (LDA)
• Case Study: PCA versus LDA
− A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on
Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp. 228-
233, 2001.
• Is LDA always better than PCA?
− There has been a tendency in the computer vision community to
prefer LDA over PCA.
− This is mainly because LDA deals directly with discrimination
between classes while PCA does not pay attention to the underlying
class structure.
− Main results of this study:
(1) When the training set is small, PCA can outperform LDA.
(2) When the number of samples is large and representative for each
class, LDA outperforms PCA.
54
Linear Discriminant Analysis (LDA)
• Is LDA always better than PCA? – cont.
55
Linear Discriminant Analysis (LDA)
• Is LDA always better than PCA? – cont.
LDA is not always better when training set is small
56
Linear Discriminant Analysis (LDA)
• Is LDA always better than PCA? – cont.
LDA outperforms PCA when training set is large
END
57

Contenu connexe

Tendances

Feature Extraction and Principal Component Analysis
Feature Extraction and Principal Component AnalysisFeature Extraction and Principal Component Analysis
Feature Extraction and Principal Component AnalysisSayed Abulhasan Quadri
 
Neural Networks and Deep Learning Basics
Neural Networks and Deep Learning BasicsNeural Networks and Deep Learning Basics
Neural Networks and Deep Learning BasicsJon Lederman
 
Edge Detection and Segmentation
Edge Detection and SegmentationEdge Detection and Segmentation
Edge Detection and SegmentationA B Shinde
 
Linear Discriminant Analysis (LDA)
Linear Discriminant Analysis (LDA)Linear Discriminant Analysis (LDA)
Linear Discriminant Analysis (LDA)Anmol Dwivedi
 
Support Vector Machines (SVM)
Support Vector Machines (SVM)Support Vector Machines (SVM)
Support Vector Machines (SVM)FAO
 
Cross validation
Cross validationCross validation
Cross validationRidhaAfrawe
 
Implementing Minimum Error Rate Classifier
Implementing Minimum Error Rate ClassifierImplementing Minimum Error Rate Classifier
Implementing Minimum Error Rate ClassifierDipesh Shome
 
Introduction to Linear Discriminant Analysis
Introduction to Linear Discriminant AnalysisIntroduction to Linear Discriminant Analysis
Introduction to Linear Discriminant AnalysisJaclyn Kokx
 
backpropagation in neural networks
backpropagation in neural networksbackpropagation in neural networks
backpropagation in neural networksAkash Goel
 
A brief introduction to mutual information and its application
A brief introduction to mutual information and its applicationA brief introduction to mutual information and its application
A brief introduction to mutual information and its applicationHyun-hwan Jeong
 
Support Vector Machines
Support Vector MachinesSupport Vector Machines
Support Vector Machinesnextlib
 
Object Detection using Deep Neural Networks
Object Detection using Deep Neural NetworksObject Detection using Deep Neural Networks
Object Detection using Deep Neural NetworksUsman Qayyum
 
Semi-Supervised Learning
Semi-Supervised LearningSemi-Supervised Learning
Semi-Supervised LearningLukas Tencer
 
Hyperparameter Optimization for Machine Learning
Hyperparameter Optimization for Machine LearningHyperparameter Optimization for Machine Learning
Hyperparameter Optimization for Machine LearningFrancesco Casalegno
 
3 - Finding similar items
3 - Finding similar items3 - Finding similar items
3 - Finding similar itemsViet-Trung TRAN
 
Dimensionality Reduction
Dimensionality ReductionDimensionality Reduction
Dimensionality ReductionSaad Elbeleidy
 
ImageProcessing10-Segmentation(Thresholding) (1).ppt
ImageProcessing10-Segmentation(Thresholding) (1).pptImageProcessing10-Segmentation(Thresholding) (1).ppt
ImageProcessing10-Segmentation(Thresholding) (1).pptVikramBarapatre2
 

Tendances (20)

Self-organizing map
Self-organizing mapSelf-organizing map
Self-organizing map
 
Feature Extraction and Principal Component Analysis
Feature Extraction and Principal Component AnalysisFeature Extraction and Principal Component Analysis
Feature Extraction and Principal Component Analysis
 
Neural Networks and Deep Learning Basics
Neural Networks and Deep Learning BasicsNeural Networks and Deep Learning Basics
Neural Networks and Deep Learning Basics
 
04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks
 
Edge Detection and Segmentation
Edge Detection and SegmentationEdge Detection and Segmentation
Edge Detection and Segmentation
 
Linear Discriminant Analysis (LDA)
Linear Discriminant Analysis (LDA)Linear Discriminant Analysis (LDA)
Linear Discriminant Analysis (LDA)
 
Decision trees
Decision treesDecision trees
Decision trees
 
Support Vector Machines (SVM)
Support Vector Machines (SVM)Support Vector Machines (SVM)
Support Vector Machines (SVM)
 
Cross validation
Cross validationCross validation
Cross validation
 
Implementing Minimum Error Rate Classifier
Implementing Minimum Error Rate ClassifierImplementing Minimum Error Rate Classifier
Implementing Minimum Error Rate Classifier
 
Introduction to Linear Discriminant Analysis
Introduction to Linear Discriminant AnalysisIntroduction to Linear Discriminant Analysis
Introduction to Linear Discriminant Analysis
 
backpropagation in neural networks
backpropagation in neural networksbackpropagation in neural networks
backpropagation in neural networks
 
A brief introduction to mutual information and its application
A brief introduction to mutual information and its applicationA brief introduction to mutual information and its application
A brief introduction to mutual information and its application
 
Support Vector Machines
Support Vector MachinesSupport Vector Machines
Support Vector Machines
 
Object Detection using Deep Neural Networks
Object Detection using Deep Neural NetworksObject Detection using Deep Neural Networks
Object Detection using Deep Neural Networks
 
Semi-Supervised Learning
Semi-Supervised LearningSemi-Supervised Learning
Semi-Supervised Learning
 
Hyperparameter Optimization for Machine Learning
Hyperparameter Optimization for Machine LearningHyperparameter Optimization for Machine Learning
Hyperparameter Optimization for Machine Learning
 
3 - Finding similar items
3 - Finding similar items3 - Finding similar items
3 - Finding similar items
 
Dimensionality Reduction
Dimensionality ReductionDimensionality Reduction
Dimensionality Reduction
 
ImageProcessing10-Segmentation(Thresholding) (1).ppt
ImageProcessing10-Segmentation(Thresholding) (1).pptImageProcessing10-Segmentation(Thresholding) (1).ppt
ImageProcessing10-Segmentation(Thresholding) (1).ppt
 

En vedette

Linear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its GeneralizationLinear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its Generalization일상 온
 
face recognition system
face recognition systemface recognition system
face recognition systemAnil Kumar
 
LDA Beginner's Tutorial
LDA Beginner's TutorialLDA Beginner's Tutorial
LDA Beginner's TutorialWayne Lee
 
Some contributions to PCA for time series
Some contributions to PCA for time seriesSome contributions to PCA for time series
Some contributions to PCA for time seriesIsabel Magalhães
 
Latent Dirichlet Allocation
Latent Dirichlet AllocationLatent Dirichlet Allocation
Latent Dirichlet AllocationMarco Righini
 
Topic Models, LDA and all that
Topic Models, LDA and all thatTopic Models, LDA and all that
Topic Models, LDA and all thatZhibo Xiao
 
LDA presentation
LDA presentationLDA presentation
LDA presentationMohit Gupta
 
Text Mining using LDA with Context
Text Mining using LDA with ContextText Mining using LDA with Context
Text Mining using LDA with ContextSteffen Staab
 
PCA Based Face Recognition System
PCA Based Face Recognition SystemPCA Based Face Recognition System
PCA Based Face Recognition SystemMd. Atiqur Rahman
 
Face recognition technology - BEST PPT
Face recognition technology - BEST PPTFace recognition technology - BEST PPT
Face recognition technology - BEST PPTSiddharth Modi
 
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vecword2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec👋 Christopher Moody
 
2012 mdsp pr09 pca lda
2012 mdsp pr09 pca lda2012 mdsp pr09 pca lda
2012 mdsp pr09 pca ldanozomuhamada
 

En vedette (20)

PCA vs LDA
PCA vs LDAPCA vs LDA
PCA vs LDA
 
Face recognition using LDA
Face recognition using LDAFace recognition using LDA
Face recognition using LDA
 
Linear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its GeneralizationLinear Discriminant Analysis and Its Generalization
Linear Discriminant Analysis and Its Generalization
 
face recognition system
face recognition systemface recognition system
face recognition system
 
Lda
LdaLda
Lda
 
LDA Beginner's Tutorial
LDA Beginner's TutorialLDA Beginner's Tutorial
LDA Beginner's Tutorial
 
Pca ppt
Pca pptPca ppt
Pca ppt
 
Some contributions to PCA for time series
Some contributions to PCA for time seriesSome contributions to PCA for time series
Some contributions to PCA for time series
 
LDA
LDALDA
LDA
 
Latent Dirichlet Allocation
Latent Dirichlet AllocationLatent Dirichlet Allocation
Latent Dirichlet Allocation
 
Topic Models, LDA and all that
Topic Models, LDA and all thatTopic Models, LDA and all that
Topic Models, LDA and all that
 
LDA presentation
LDA presentationLDA presentation
LDA presentation
 
Text Mining using LDA with Context
Text Mining using LDA with ContextText Mining using LDA with Context
Text Mining using LDA with Context
 
Principal component analysis
Principal component analysisPrincipal component analysis
Principal component analysis
 
PCA Based Face Recognition System
PCA Based Face Recognition SystemPCA Based Face Recognition System
PCA Based Face Recognition System
 
Principal Component Analysis
Principal Component AnalysisPrincipal Component Analysis
Principal Component Analysis
 
07. PCA
07. PCA07. PCA
07. PCA
 
Face recognition technology - BEST PPT
Face recognition technology - BEST PPTFace recognition technology - BEST PPT
Face recognition technology - BEST PPT
 
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vecword2vec, LDA, and introducing a new hybrid algorithm: lda2vec
word2vec, LDA, and introducing a new hybrid algorithm: lda2vec
 
2012 mdsp pr09 pca lda
2012 mdsp pr09 pca lda2012 mdsp pr09 pca lda
2012 mdsp pr09 pca lda
 

Similaire à Understandig PCA and LDA

Fr pca lda
Fr pca ldaFr pca lda
Fr pca ldaultraraj
 
DimensionalityReduction.pptx
DimensionalityReduction.pptxDimensionalityReduction.pptx
DimensionalityReduction.pptx36rajneekant
 
Dimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptxDimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptxSivam Chinna
 
pcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptxpcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptxABINASHPADHY6
 
introduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptxintroduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptxPratik Gohel
 
Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Mostafa G. M. Mostafa
 
Matrix Factorization In Recommender Systems
Matrix Factorization In Recommender SystemsMatrix Factorization In Recommender Systems
Matrix Factorization In Recommender SystemsYONG ZHENG
 
5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdf5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdfRahul926331
 
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image DescriptorsPCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image Descriptorswolf
 
Singular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptxSingular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptxrajalakshmi5921
 
EDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptxEDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptxrajalakshmi5921
 
The following ppt is about principal component analysis
The following ppt is about principal component analysisThe following ppt is about principal component analysis
The following ppt is about principal component analysisSushmit8
 
Computational Giants_nhom.pptx
Computational Giants_nhom.pptxComputational Giants_nhom.pptx
Computational Giants_nhom.pptxThAnhonc
 
Lect4 principal component analysis-I
Lect4 principal component analysis-ILect4 principal component analysis-I
Lect4 principal component analysis-Ihktripathy
 
Face Recognition Using Neural Networks
Face Recognition Using Neural NetworksFace Recognition Using Neural Networks
Face Recognition Using Neural NetworksCSCJournals
 
Aaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reductionAaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reductionAminaRepo
 

Similaire à Understandig PCA and LDA (20)

Fr pca lda
Fr pca ldaFr pca lda
Fr pca lda
 
DimensionalityReduction.pptx
DimensionalityReduction.pptxDimensionalityReduction.pptx
DimensionalityReduction.pptx
 
Dimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptxDimensionality Reduction and feature extraction.pptx
Dimensionality Reduction and feature extraction.pptx
 
pcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptxpcappt-140121072949-phpapp01.pptx
pcappt-140121072949-phpapp01.pptx
 
introduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptxintroduction to machine learning 3c-feature-extraction.pptx
introduction to machine learning 3c-feature-extraction.pptx
 
Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)Neural Networks: Principal Component Analysis (PCA)
Neural Networks: Principal Component Analysis (PCA)
 
Matrix Factorization In Recommender Systems
Matrix Factorization In Recommender SystemsMatrix Factorization In Recommender Systems
Matrix Factorization In Recommender Systems
 
5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdf5 DimensionalityReduction.pdf
5 DimensionalityReduction.pdf
 
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image DescriptorsPCA-SIFT: A More Distinctive Representation for Local Image Descriptors
PCA-SIFT: A More Distinctive Representation for Local Image Descriptors
 
Singular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptxSingular Value Decomposition (SVD).pptx
Singular Value Decomposition (SVD).pptx
 
EDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptxEDAB Module 5 Singular Value Decomposition (SVD).pptx
EDAB Module 5 Singular Value Decomposition (SVD).pptx
 
pca.ppt
pca.pptpca.ppt
pca.ppt
 
The following ppt is about principal component analysis
The following ppt is about principal component analysisThe following ppt is about principal component analysis
The following ppt is about principal component analysis
 
PCA Final.pptx
PCA Final.pptxPCA Final.pptx
PCA Final.pptx
 
Fa18_P2.pptx
Fa18_P2.pptxFa18_P2.pptx
Fa18_P2.pptx
 
Machine learning meetup
Machine learning meetupMachine learning meetup
Machine learning meetup
 
Computational Giants_nhom.pptx
Computational Giants_nhom.pptxComputational Giants_nhom.pptx
Computational Giants_nhom.pptx
 
Lect4 principal component analysis-I
Lect4 principal component analysis-ILect4 principal component analysis-I
Lect4 principal component analysis-I
 
Face Recognition Using Neural Networks
Face Recognition Using Neural NetworksFace Recognition Using Neural Networks
Face Recognition Using Neural Networks
 
Aaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reductionAaa ped-17-Unsupervised Learning: Dimensionality reduction
Aaa ped-17-Unsupervised Learning: Dimensionality reduction
 

Plus de Dr. Syed Hassan Amin

Greenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparisonGreenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparisonDr. Syed Hassan Amin
 
Multitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq RecognitionMultitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq RecognitionDr. Syed Hassan Amin
 
Thin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better CodeThin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better CodeDr. Syed Hassan Amin
 
Improving Code Quality Through Effective Review Process
Improving Code Quality Through Effective  Review ProcessImproving Code Quality Through Effective  Review Process
Improving Code Quality Through Effective Review ProcessDr. Syed Hassan Amin
 
Software Project Management Tips and Tricks
Software Project Management Tips and TricksSoftware Project Management Tips and Tricks
Software Project Management Tips and TricksDr. Syed Hassan Amin
 
Improving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design PrinciplesImproving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design PrinciplesDr. Syed Hassan Amin
 
Learning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve JobsLearning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve JobsDr. Syed Hassan Amin
 
Understanding and Managing Technical Debt
Understanding and Managing Technical DebtUnderstanding and Managing Technical Debt
Understanding and Managing Technical DebtDr. Syed Hassan Amin
 
An OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq FontAn OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq FontDr. Syed Hassan Amin
 

Plus de Dr. Syed Hassan Amin (12)

Greenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparisonGreenplum versus redshift and actian vectorwise comparison
Greenplum versus redshift and actian vectorwise comparison
 
Introduction To Docker
Introduction To  DockerIntroduction To  Docker
Introduction To Docker
 
Laravel Unit Testing
Laravel Unit TestingLaravel Unit Testing
Laravel Unit Testing
 
Multitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq RecognitionMultitier holistic Approach for urdu Nastaliq Recognition
Multitier holistic Approach for urdu Nastaliq Recognition
 
Agile Scrum Methodology
Agile Scrum MethodologyAgile Scrum Methodology
Agile Scrum Methodology
 
Thin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better CodeThin Controllers Fat Models - How to Write Better Code
Thin Controllers Fat Models - How to Write Better Code
 
Improving Code Quality Through Effective Review Process
Improving Code Quality Through Effective  Review ProcessImproving Code Quality Through Effective  Review Process
Improving Code Quality Through Effective Review Process
 
Software Project Management Tips and Tricks
Software Project Management Tips and TricksSoftware Project Management Tips and Tricks
Software Project Management Tips and Tricks
 
Improving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design PrinciplesImproving Software Quality Using Object Oriented Design Principles
Improving Software Quality Using Object Oriented Design Principles
 
Learning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve JobsLearning Technology Leadership from Steve Jobs
Learning Technology Leadership from Steve Jobs
 
Understanding and Managing Technical Debt
Understanding and Managing Technical DebtUnderstanding and Managing Technical Debt
Understanding and Managing Technical Debt
 
An OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq FontAn OCR System for recognition of Urdu text in Nastaliq Font
An OCR System for recognition of Urdu text in Nastaliq Font
 

Dernier

Market Analysis in the 5 Largest Economic Countries in Southeast Asia.pdf
Market Analysis in the 5 Largest Economic Countries in Southeast Asia.pdfMarket Analysis in the 5 Largest Economic Countries in Southeast Asia.pdf
Market Analysis in the 5 Largest Economic Countries in Southeast Asia.pdfRachmat Ramadhan H
 
VidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptxVidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptxolyaivanovalion
 
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service Bhilai
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service BhilaiLow Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service Bhilai
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service BhilaiSuhani Kapoor
 
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptx
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptxBPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptx
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptxMohammedJunaid861692
 
Schema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfSchema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfLars Albertsson
 
Week-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionWeek-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionfulawalesam
 
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...Suhani Kapoor
 
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改atducpo
 
Introduction-to-Machine-Learning (1).pptx
Introduction-to-Machine-Learning (1).pptxIntroduction-to-Machine-Learning (1).pptx
Introduction-to-Machine-Learning (1).pptxfirstjob4
 
Industrialised data - the key to AI success.pdf
Industrialised data - the key to AI success.pdfIndustrialised data - the key to AI success.pdf
Industrialised data - the key to AI success.pdfLars Albertsson
 
Brighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingBrighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingNeil Barnes
 
Smarteg dropshipping via API with DroFx.pptx
Smarteg dropshipping via API with DroFx.pptxSmarteg dropshipping via API with DroFx.pptx
Smarteg dropshipping via API with DroFx.pptxolyaivanovalion
 
RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998YohFuh
 
100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptxAnupama Kate
 
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Serviceranjana rawat
 
Edukaciniai dropshipping via API with DroFx
Edukaciniai dropshipping via API with DroFxEdukaciniai dropshipping via API with DroFx
Edukaciniai dropshipping via API with DroFxolyaivanovalion
 
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...Suhani Kapoor
 
Delhi Call Girls CP 9711199171 ☎✔👌✔ Whatsapp Hard And Sexy Vip Call
Delhi Call Girls CP 9711199171 ☎✔👌✔ Whatsapp Hard And Sexy Vip CallDelhi Call Girls CP 9711199171 ☎✔👌✔ Whatsapp Hard And Sexy Vip Call
Delhi Call Girls CP 9711199171 ☎✔👌✔ Whatsapp Hard And Sexy Vip Callshivangimorya083
 
Ravak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptxRavak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptxolyaivanovalion
 

Dernier (20)

Market Analysis in the 5 Largest Economic Countries in Southeast Asia.pdf
Market Analysis in the 5 Largest Economic Countries in Southeast Asia.pdfMarket Analysis in the 5 Largest Economic Countries in Southeast Asia.pdf
Market Analysis in the 5 Largest Economic Countries in Southeast Asia.pdf
 
VidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptxVidaXL dropshipping via API with DroFx.pptx
VidaXL dropshipping via API with DroFx.pptx
 
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service Bhilai
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service BhilaiLow Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service Bhilai
Low Rate Call Girls Bhilai Anika 8250192130 Independent Escort Service Bhilai
 
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptx
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptxBPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptx
BPAC WITH UFSBI GENERAL PRESENTATION 18_05_2017-1.pptx
 
Schema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdfSchema on read is obsolete. Welcome metaprogramming..pdf
Schema on read is obsolete. Welcome metaprogramming..pdf
 
Week-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interactionWeek-01-2.ppt BBB human Computer interaction
Week-01-2.ppt BBB human Computer interaction
 
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
VIP High Class Call Girls Jamshedpur Anushka 8250192130 Independent Escort Se...
 
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
代办国外大学文凭《原版美国UCLA文凭证书》加州大学洛杉矶分校毕业证制作成绩单修改
 
Introduction-to-Machine-Learning (1).pptx
Introduction-to-Machine-Learning (1).pptxIntroduction-to-Machine-Learning (1).pptx
Introduction-to-Machine-Learning (1).pptx
 
Industrialised data - the key to AI success.pdf
Industrialised data - the key to AI success.pdfIndustrialised data - the key to AI success.pdf
Industrialised data - the key to AI success.pdf
 
Brighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data StorytellingBrighton SEO | April 2024 | Data Storytelling
Brighton SEO | April 2024 | Data Storytelling
 
Smarteg dropshipping via API with DroFx.pptx
Smarteg dropshipping via API with DroFx.pptxSmarteg dropshipping via API with DroFx.pptx
Smarteg dropshipping via API with DroFx.pptx
 
RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998RA-11058_IRR-COMPRESS Do 198 series of 1998
RA-11058_IRR-COMPRESS Do 198 series of 1998
 
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
VIP Call Girls Service Charbagh { Lucknow Call Girls Service 9548273370 } Boo...
 
100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx100-Concepts-of-AI by Anupama Kate .pptx
100-Concepts-of-AI by Anupama Kate .pptx
 
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
(PARI) Call Girls Wanowrie ( 7001035870 ) HI-Fi Pune Escorts Service
 
Edukaciniai dropshipping via API with DroFx
Edukaciniai dropshipping via API with DroFxEdukaciniai dropshipping via API with DroFx
Edukaciniai dropshipping via API with DroFx
 
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
VIP High Profile Call Girls Amravati Aarushi 8250192130 Independent Escort Se...
 
Delhi Call Girls CP 9711199171 ☎✔👌✔ Whatsapp Hard And Sexy Vip Call
Delhi Call Girls CP 9711199171 ☎✔👌✔ Whatsapp Hard And Sexy Vip CallDelhi Call Girls CP 9711199171 ☎✔👌✔ Whatsapp Hard And Sexy Vip Call
Delhi Call Girls CP 9711199171 ☎✔👌✔ Whatsapp Hard And Sexy Vip Call
 
Ravak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptxRavak dropshipping via API with DroFx.pptx
Ravak dropshipping via API with DroFx.pptx
 

Understandig PCA and LDA

  • 1. CS 312 Character and Pattern Recognition – Spring 2009 Dimensionality Reduction Using PCA/LDA Chapter 3 (Duda et al.) – Section 3.8 Thanks to Prof. Bebis Case Studies: Face Recognition Using Dimensionality Reduction M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive Neuroscience, 3(1), pp. 71-86, 1991. D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image Retrieval", IEEE Transactions on Pattern Analysis and Machine Intelligence, 18(8), pp. 831-836, 1996. A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp. 228-233, 2001.
  • 2. 2 Dimensionality Reduction • One approach to deal with high dimensional data is by reducing their dimensionality. • Project high dimensional data onto a lower dimensional sub- space using linear or non-linear transformations.
  • 3. 3 Dimensionality Reduction • Linear transformations are simple to compute and tractable. • Classical –linear- approaches: – Principal Component Analysis (PCA) – Fisher Discriminant Analysis (FDA) ( )t i i iY U X b u a= = k x 1 k x d d x 1 (k<<d)k x 1 k x d d x 1 (k<<d)
  • 4. 4 Principal Component Analysis (PCA) • Each dimensionality reduction technique finds an appropriate transformation by satisfying certain criteria (e.g., information loss, data discrimination, etc.) • The goal of PCA is to reduce the dimensionality of the data while retaining as much as possible of the variation present in the dataset.
  • 5. 5 Principal Component Analysis (PCA) 1 1 2 2 1 2 ˆ ... where , ,..., isa basein the -dimensionalsub-space(K<N) K K K x b u b u b u u u u K = + + + ˆx x= 1 1 2 2 1 2 ... where , ,..., isa basein theoriginal N-dimensionalspace N N n x a v a v a v v v v = + + + • Find a basis in a low dimensional sub-space: − Approximate vectors by projecting them in a low dimensional sub-space: (1) Original space representation: (2) Lower-dimensional sub-space representation: • Note: if K=N, then
  • 6. 6 Principal Component Analysis (PCA) • Example (K=N):
  • 7. 7 Principal Component Analysis (PCA) • Information loss − Dimensionality reduction implies information loss !! − PCA preserves as much information as possible: • What is the “best” lower dimensional sub-space? The “best” low-dimensional space is centered at the sample mean and has directions determined by the “best” eigenvectors of the covariance matrix of the data x. − By “best” eigenvectors we mean those corresponding to the largest eigenvalues ( i.e., “principal components”). − Since the covariance matrix is real and symmetric, these eigenvectors are orthogonal and form a set of basis vectors. (see pp. 114-117 in textbook for a proof) ˆmin || || (reconstruction error)x x−
  • 8. 8 Principal Component Analysis (PCA) • Methodology − Suppose x1, x2, ..., xM are N x 1 vectors
  • 9. 9 Principal Component Analysis (PCA) • Methodology – cont. ( )T i ib u x x= −
  • 10. 10 Principal Component Analysis (PCA) • Eigenvalue spectrum λi K λN
  • 11. 11 Principal Component Analysis (PCA) • Linear transformation implied by PCA − The linear transformation RN → RK that performs the dimensionality reduction is:
  • 12. 12 Principal Component Analysis (PCA) • Geometric interpretation − PCA projects the data along the directions where the data varies the most. − These directions are determined by the eigenvectors of the covariance matrix corresponding to the largest eigenvalues. − The magnitude of the eigenvalues corresponds to the variance of the data along the eigenvector directions.
  • 13. 13 Principal Component Analysis (PCA) • How many principal components to keep? − To choose K, you can use the following criterion:
  • 14. 14 Principal Component Analysis (PCA) • What is the error due to dimensionality reduction? • It can be shown that the average error due to dimensionality reduction is equal to: e
  • 15. 15 Principal Component Analysis (PCA) • Standardization − The principal components are dependent on the units used to measure the original variables as well as on the range of values they assume. − We should always standardize the data prior to using PCA. − A common standardization method is to transform all the data to have zero mean and unit standard deviation:
  • 16. 16 Principal Component Analysis (PCA) • Case Study: Eigenfaces for Face Detection/Recognition − M. Turk, A. Pentland, "Eigenfaces for Recognition", Journal of Cognitive Neuroscience, vol. 3, no. 1, pp. 71-86, 1991. • Face Recognition − The simplest approach is to think of it as a template matching problem − Problems arise when performing recognition in a high- dimensional space. − Significant improvements can be achieved by first mapping the data into a lower dimensionality space. − How to find this lower-dimensional space?
  • 17. 17 Principal Component Analysis (PCA) • Main idea behind eigenfaces ˆΦ −Ψ −Ψ average face
  • 18. 18 Principal Component Analysis (PCA) • Computation of the eigenfaces
  • 19. 19 Principal Component Analysis (PCA) • Computation of the eigenfaces – cont.
  • 20. 20 Principal Component Analysis (PCA) • Computation of the eigenfaces – cont. ui
  • 21. 21 Principal Component Analysis (PCA) • Computation of the eigenfaces – cont.
  • 22. 22 Principal Component Analysis (PCA) • Representing faces onto this basis ˆ iΦ
  • 23. 23 Principal Component Analysis (PCA) • Eigenvalue spectrum : Defined as set of Eigenvalues. • Spectrum is shown as Eigenvalues plotted in the graph shown below: λi K M λN
  • 24. 24 Principal Component Analysis (PCA) • Representing faces onto this basis – cont.
  • 25. 25 Principal Component Analysis (PCA) • Face Recognition Using Eigenfaces
  • 26. 26 Principal Component Analysis (PCA) • Face Recognition Using Eigenfaces – cont. − The distance er is called distance within the face space (difs) − Comment: we can use the common Euclidean distance to compute er, however, it has been reported that the Mahalanobis distance performs better:
  • 27. 27 Principal Component Analysis (PCA) • Face Detection Using Eigenfaces
  • 28. 28 Principal Component Analysis (PCA) • Face Detection Using Eigenfaces – cont.
  • 29. 29 Principal Component Analysis (PCA) • Reconstruction of faces and non-faces
  • 30. 30 Principal Component Analysis (PCA) • Applications − Face detection, tracking, and recognition dffs
  • 31. 31 Principal Component Analysis (PCA) • Problems in face recognition: − Background (de-emphasize the outside of the face – e.g., by multiplying the input image by a 2D Gaussian window centered on the face) − Lighting conditions (performance degrades with light changes) − Scale (performance decreases quickly with changes to head size) − multi-scale eigenspaces − scale input image to multiple sizes − Orientation (performance decreases but not as fast as with scale changes) − plane rotations can be handled − out-of-plane rotations are more difficult to handle
  • 32. 32 Principal Component Analysis (PCA) • Dataset − 16 subjects − 3 orientations, 3 sizes − 3 lighting conditions, 6 resolutions (512x512 ... 16x16) − Total number of images: 2,592
  • 33. 33 Principal Component Analysis (PCA) • Experiment 1 − Used various sets of 16 images for training − One image/person, taken under the same conditions − Classify the rest images as one of the 16 individuals − 7 eigenfaces were used − No rejections (i.e., no threshold on difs) − Performed a large number of experiments and averaged the results: − 96% correct averaged over light variation − 85% correct averaged over orientation variation − 64% correct averaged over size variation
  • 34. 34 Principal Component Analysis (PCA) • Experiment 2 − Considered rejections (i.e., by thresholding difs) − There is a tradeoff between correct recognition and rejections. − Adjusting the threshold to achieve 100% recognition accuracy resulted in: − 19% rejections while varying lighting − 39% rejections while varying orientation − 60% rejections while varying size • Experiment 3 − Reconstruction using partial information
  • 35. 35 Principal Component Analysis (PCA) • PCA and classification − PCA is not always an optimal dimensionality-reduction procedure for classification purposes. • Multiple classes and PCA − Suppose there are C classes in the training data. − PCA is based on the sample covariance which characterizes the scatter of the entire data set, irrespective of class-membership. − The projection axes chosen by PCA might not provide good discrimination power.
  • 36. 36 Principal Component Analysis (PCA) • PCA and classification (cont’d)
  • 37. 37 Linear Discriminant Analysis (LDA) • What is the goal of LDA? − Perform dimensionality reduction “while preserving as much of the class discriminatory information as possible”. − Seeks to find directions along which the classes are best separated. − Takes into consideration the scatter within-classes but also the scatter between-classes. − More capable of distinguishing image variation due to identity from variation due to other sources such as illumination and expression.
  • 39. 39 Linear Discriminant Analysis (LDA) • Notation (Sb has at most rank C-1) (each sub-matrix has rank 1 or less, i.e., outer product of two vectors) 1 1 ( )( ) iMC T w j i j i i j S x x = = = − −∑∑ μ μ
  • 40. 40 Linear Discriminant Analysis (LDA) • Methodology − LDA computes a transformation that maximizes the between-class scatter while minimizing the within-class scatter: | | | | max max | | | | T b b T w w U S U S U S U S = % % T U=y x products of eigenvalues ! projection matrix ,b wS S% % : scatter matrices of the projected data y
  • 41. 41 Linear Discriminant Analysis (LDA) • Linear transformation implied by LDA − The linear transformation is given by a matrix U whose columns are the eigenvectors of the above problem (i.e., called Fisherfaces). − The LDA solution is given by the eigenvectors of the generalized eigenvector problem: − Important: Since Sb has at most rank C-1, the max number of eigenvectors with non-zero eigenvalues is C-1 (i.e., max dimensionality of sub-space is C-1)
  • 42. 42 Linear Discriminant Analysis (LDA) • Does Sw -1 always exist? − If Sw is non-singular, we can obtain a conventional eigenvalue problem by writing: − In practice, Sw is often singular since the data are image vectors with large dimensionality while the size of the data set is much smaller (M << N )
  • 43. 43 Linear Discriminant Analysis (LDA) • Does Sw -1 always exist? – cont. − To alleviate this problem, we can use PCA first: 1) PCA is first applied to the data set to reduce its dimensionality. 2) LDA is then applied to find the most discriminative directions:
  • 44. 44 Linear Discriminant Analysis (LDA) • Case Study: Using Discriminant Eigenfeatures for Image Retrieval − D. Swets, J. Weng, "Using Discriminant Eigenfeatures for Image Retrieval", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 18, no. 8, pp. 831-836, 1996. • Content-based image retrieval − The application being studied here is query-by-example image retrieval. − The paper deals with the problem of selecting a good set of image features for content-based image retrieval.
  • 45. 45 Linear Discriminant Analysis (LDA) • Assumptions − "Well-framed" images are required as input for training and query- by-example test probes. − Only a small variation in the size, position, and orientation of the objects in the images is allowed.
  • 46. 46 Linear Discriminant Analysis (LDA) • Some terminology − Most Expressive Features (MEF): the features (projections) obtained using PCA. − Most Discriminating Features (MDF): the features (projections) obtained using LDA. • Numerical problems − When computing the eigenvalues/eigenvectors of Sw -1 SBuk = λkuk numerically, the computations can be unstable since Sw -1 SB is not always symmetric. − See paper for a way to find the eigenvalues/eigenvectors in a stable way.
  • 47. 47 Linear Discriminant Analysis (LDA) • Factors unrelated to classification − MEF vectors show the tendency of PCA to capture major variations in the training set such as lighting direction. − MDF vectors discount those factors unrelated to classification.
  • 48. 48 Linear Discriminant Analysis (LDA) • Clustering effect 1) Generate the set of MEFs/MDFs for each image in the training set. 2) Given an query image, compute its MEFs/MDFs using the same procedure. 3) Find the k closest neighbors for retrieval (e.g., using Euclidean distance). • Methodology
  • 49. 49 Linear Discriminant Analysis (LDA) • Experiments and results − Face images − A set of face images was used with 2 expressions, 3 lighting conditions. − Testing was performed using a disjoint set of images: • One image, randomly chosen, from each individual.
  • 51. 51 Linear Discriminant Analysis (LDA) − Examples of correct search probes
  • 52. 52 Linear Discriminant Analysis (LDA) − Example of a failed search probe
  • 53. 53 Linear Discriminant Analysis (LDA) • Case Study: PCA versus LDA − A. Martinez, A. Kak, "PCA versus LDA", IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 23, no. 2, pp. 228- 233, 2001. • Is LDA always better than PCA? − There has been a tendency in the computer vision community to prefer LDA over PCA. − This is mainly because LDA deals directly with discrimination between classes while PCA does not pay attention to the underlying class structure. − Main results of this study: (1) When the training set is small, PCA can outperform LDA. (2) When the number of samples is large and representative for each class, LDA outperforms PCA.
  • 54. 54 Linear Discriminant Analysis (LDA) • Is LDA always better than PCA? – cont.
  • 55. 55 Linear Discriminant Analysis (LDA) • Is LDA always better than PCA? – cont. LDA is not always better when training set is small
  • 56. 56 Linear Discriminant Analysis (LDA) • Is LDA always better than PCA? – cont. LDA outperforms PCA when training set is large

Notes de l'éditeur

  1. Good afternoon and thank you everyone for coming. My talk today will describe the research I performed at the IRIS at USC, the object of this work being to build a computational framework that addresses the problem of motion analysis and interpretation.