Machine Vision and Applications (2009) 20:35–46
DOI 10.1007/s00138-007-0103-1
ORIGINAL PAPER
Complete discriminant evaluation and feature extraction
in kernel space for face recognition
Xudong Jiang · Bappaditya Mandal · Alex Kot
Received: 20 December 2006 / Accepted: 21 July 2007 / Published online: 15 November 2007
© Springer-Verlag 2007
Abstract This work proposes a method to decompose the
kernel within-class eigenspace into two subspaces: a reliable
subspace spanned mainly by the facial variation and an unre-
liable subspace due to limited number of training samples.
A weighting function is proposed to circumvent undue scal-
ing of eigenvectors corresponding to the unreliable small and
zero eigenvalues. Eigenfeatures are then extracted by the dis-
criminant evaluation in the whole kernel space. These efforts
facilitate a discriminative and stable low-dimensional fea-
ture representation of the face image. Experimental results
on FERET, ORL and GT databases show that our approach
consistently outperforms other kernel based face recognition
methods.
Keywords Face recognition · Kernel discriminant
analysis · Feature extraction · Subspace methods
1 Introduction
Face recognition has drawn considerable attention in bio-
metric society because of its property of non-intrusiveness,
which is recognizing a person from a distance. Numerous
linear and nonlinear subspace based face recognition
methods are proposed in the last two decades [25, 30, 35].
The most popular nonlinear methods are kernel principal
X. Jiang (B ) · B. Mandal · A. Kot
School of Electrical and Electronic Engineering,
Nanyang Technological University,
Singapore 639798, Singapore
e-mail: exdjiang@ntu.edu.sg
B. Mandal
e-mail: bapp0001@ntu.edu.sg
A. Kot
e-mail: eackot@ntu.edu.sg
component analysis (KPCA) [29] and kernel Fisher discrim-
inant analysis (KFDA) [22]. These kernel based methods
and their variations encode pattern information based on
higher order dependencies and are tactful to the dependen-
cies among pixels in the samples. The kernel mappings can
capture the nonlinearities and complex relationships among
the input data that exist due to the expression, illumination
and pose variations.
The basic idea of these kernel methods is to apply nonlin-
ear mapping Φ : X ∈ R
n
→ Φ( X ) ∈ H in the image space
R
n
, followed by linear subspace methods like PCA and FDA
in the mapped feature space H. Examples include KPCA [29]
and KFDA [22, 23]. Since the feature space H can be very
high or possibly infinite dimensional and the orthogonality
needs to be characterized in such a space, it is reasonable to
view H as a Hilbert space. It is difficult to compute the dot
products in the high dimensional feature space H. Instead of
mapping the data explicitly, the feature space can be com-
puted by using the kernel trick, in which the inner products
〈Φ( X
ij
),Φ( X
st
)〉 in H can be replaced with a kernel function
K ( X
ij
, X
st
), where K ( X
ij
, X
st
) =〈Φ( X
ij
),Φ( X
st
)〉 and
X
ij
, X
st
are sample vectors in the image space R
n
. So, the
nonlinear mapping Φ can be performed implicitly in image
space R
n
[28, 31]. Numerous studies [8, 30, 34] demonstrate
that these kernel based approaches are effective in some
real-world applications. However, the basic subspace analy-
sis has still outstanding challenging problems when applied
to the face recognition due to the high dimensionality of
the face image and the finite number of training samples in
practice.
Most of the kernel subspace based face recognition meth-
ods perform dimensionality reduction or discard a subspace
before the discriminant evaluation. A popular method called
kernel Fisherface [34] applies PCA first for dimensional-
ity reduction so as to make the within-class scatter matrix
123