Research Letters in Signal Processing

Volume 2008 (2008), Article ID 486247, 5 pages

http://dx.doi.org/10.1155/2008/486247

## Extraction and Recognition of Nonlinear Interval-Type Features Using Symbolic KDA Algorithm with Application to Face Recognition

^{1}Department of Studies in Computer Science, Gulbarga University, Gulbarga, 585-106 Karnataka, India^{2}Department of Studies in Computer Science, Kuvempu University, Shankaraghatta, 577-451 Karnataka, India

Received 21 July 2007; Accepted 12 February 2008

Academic Editor: Paul Cristea

Copyright © 2008 P. S. Hiremath and C. J. Prabhakar. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

We present symbolic kernel discriminant analysis (symbolic KDA) for face recognition in the framework of symbolic data analysis. Classical KDA extracts features, which are single-valued in nature to represent face images. These single-valued variables may not be able to capture variation of each feature in all the images of same subject; this leads to loss of information. The symbolic KDA algorithm extracts most discriminating nonlinear interval-type features which optimally discriminate among the classes represented in the training set. The proposed method has been successfully tested for face recognition using two databases, ORL database and Yale face database. The effectiveness of the proposed method is shown in terms of comparative performance against popular face recognition methods such as kernel Eigenface method and kernel Fisherface method. Experimental results show that symbolic KDA yields improved recognition rate.

#### 1. Introduction

Kernel principal component analysis (KPCA) [1, 2] and kernel Fisher discriminant analysis (KFD) [3] have aroused considerable interest in the face recogniation problem. KPCA was originally developed by Schölkopf et al., in 1998, while KFD was first proposed by Mika et al., in 1999 [3]. Subsequent research saw the development of series of KFD algorithms. The KFD based algorithms use the pixel intensity values in a face image as the features for face recognition. The pixel intensities that are used as features are represented by single valued variables. However, in many situations same face is captured in different orientation, lighting, expression and background, which all lead to image variations. The pixel intensities do change because of image variations. The use of single valued variables may not be able to capture the variation of feature values of the images of the same subject. In such a case, we need to consider the symbolic data analysis (SDA) [4–7], in which the interval-valued data are analyzed.

In this paper, new appearance based method is proposed in the framework of symbolic data analysis, namely, symbolic KDA for face recognition, which is a generalization of the classical KDA to symbolic objects. In the first step, we represent the face images as symbolic objects (symbolic faces) of interval type variables. Each symbolic face summarizes the variation of feature values through the different images of the same subject. It also drastically reduces the dimension of the image space without losing a significant amount of information.

In the second step, we apply symbolic KDA algorithm to extract interval type nonlinear discriminating features. According to this algorithm, in the first phase, we apply kernel function to symbolic faces, as a result a pattern in the original input space is mapped into a potentially much higher dimensional feature vector in the feature space, and then performs in the feature space to choose subspace dimension carefully. In the second phase, we extract interval type nonlinear discriminating features, which are robust to variations due to illumination, orientation and facial expression. Finally, minimum distance classifier with symbolic dissimilarity measure [4] is employed for classification.

The remainder of this paper is organized as follows: In Section 2, the idea of constructing the symbolic faces is given. Symbolic KDA is developed in Section 3. In Section 4, the experiments are performed on the ORL and Yale face database whereby the proposed algorithm is evaluated and compared with other methods. Finally, conclusion and discussion are given in Section 5.

#### 2. Construction of Symbolic Faces

Let be the collection of *n* face
images of the database, each of size ,
which are first-order objects. Each object , is described by a
feature vector of length where each component is a single-valued variable representing the
intensity values of the face image .
An image set is a collection of face images of *m* different subjects (face classes) and each subject has different
images with varying expressions, orientations, and illuminations. Thus there
are *m* number of second-order objects
(face classes) denoted by ,
each consisting of different individual images, ,
of a subject. We have assumed that images are belonging to a face class and are arranged
from right-side view to left-side view. The view range of each face class is
partitioned into subface classes and each subface class contains number of
images. The feature vector of th subface class of th face class ,
where is described by a vector of *p* interval variables ,
and is of length .
The interval variable of th subface class of th face class is described as where and are minimum and maximum intensity values,
respectively, among th pixels of all the images of subface class .
This interval incorporates variability of th feature inside the th subface class .

We denote The vector of interval variables is recorded for th subface class of th face class. This vector is called as *symbolic face* and is represented as
where and ;

We represent the symbolic faces by a matrix of size , consisting of column vectors ,

#### 3. Acquiring Nonlinear Subspace Using Symbolic KDA Method

Let us consider the matrix containing *qm* symbolic faces pertaining to the given set of images belonging to *m* face classes. The centers of the intervals are given by
where , and .

The data matrix containing the centers of the intervals for *qm* symbolic
faces. The *p-*dimensional vectors , and represent the centers, lower bounds, and upper bounds of the *qm* symbolic faces ,
respectively.

Let be a nonlinear mapping between the input space and the feature space.

The nonlinear mapping, usually defines a kernel function. Let define a kernel matrix by means of dot product
in the feature space:
In general, the
Fisher criterion function in the feature space *F* can be defined as
where *V* is a discriminant vector, and are the between-class scatter matrix and the within-class
scatter matrix, respectively. The between-class scatter matrix and the within-class
scatter matrix in the feature space are defined below:
where denotes the th symbolic face of th face class, is the number of training symbolic faces in
face class , is the mean of the mapped symbolic faces in
face class ,
and is the mean across all mapped symbolic faces. From the above definitions, we
have .
The discriminant vectors with respect to the Fisher criterion are actually the
eigenvectors of the generalized equation .
According to the theory of the reproducing kernel, *V* will be an expansion
of all symbolic faces in the feature space. That is, there exist coefficients such that
where and

Substituting (8)
into (6), we can obtain the following equation:
where *K* is a kernel matrix, is a matrix whose elements are From the definition of *W,* it is easy to
verify that *W* is a block matrix. In fact, it is often necessary
to find *s* discriminant vectors, denoted by ,
to extract features. Let .
The matrix *V* should satisfy the following condition:
where and .

Since, each symbolic face is located between the lower bound symbolic face and upper bound symbolic face , it is possible to find most discriminating nonlinear interval-type features .

The lower bound features of each symbolic face is given by where Similarly, the upper bound features of each symbolic face is given by Let be the test face class that contains face images of same subject with different expression, lighting condition and orientation. The test symbolic face is constructed for test face class as explained in Section 2. The lower bound test symbolic face of test symbolic face is described as . Similarly, the upper bound test symbolic face is described as .

The interval-type features of test symbolic face are computed as: where

#### 4. Experimental Results

The proposed symbolic KDA method is experimented with the face images of the ORL and Yale databases. The effectiveness of proposed method is shown in terms of comparative performance against two face recognition methods. In particular, we compared our algorithm with kernel Eigenface [8] and kernel Fisherface [9] method.

##### 4.1. Experiments Using ORL Database

We assess the feasibility and performance of the proposed symbolic KDA on the face recognition task using ORL database. The ORL face database is composed of 400 images with ten different facial views that represent various expressions and orientations for each of the 40 distinct subjects as shown in Figure 1. We have arranged images of each face class from right side view to left side view. All the 400 images from the ORL database are used to evaluate the face recognition performance of proposed method. Six images are randomly chosen from the ten images available for each subject for training, while the remaining images are used to construct the test symbolic face for each trial. We have conducted the experiments using two kernel functions namely, polynomial kernel and Gaussian kernel.

Our goal is to find appropriate kernel function and corresponding optimal kernel parameters (i.e., the order of the polynomial kernel and the width of the Gaussian kernel) for our proposed method. The experimental results shows that the order of the polynomial kernel should be three and the width of Gaussian kernel should be four for proposed symbolic KDA with respect to a minimum distance classifier.

After determining the optimal kernel parameters, we set out to select the dimension of discriminant subspace with respect to two different kernels. Table 1 shows optimal subspace for each method. The optimal parameters for each method with respect to different kernels. Besides, we find that symbolic KDA features seem more effective than features of other methods.

After selection of optimal parameters and optimal subspace for each method with respect to different kernels, all three methods are reevaluated using same set of training and testing samples. The average recognition rates for the best case are shown in Table 2. The best performance of the symbolic KDA method is better than the best performance of the kernel Eigenface and kernel Fisherface method. We note that the symbolic KDA method outperforms kernel Eigenface method and kernel Fisherface in the sense of using small number of features.

In order to examine, whether symbolic KDA is statistically significant and better than other
methods in terms of its recognition rate. We evaluate the experimental results presented in
Table 2 using McNemar’s significance test. McNemar’s test is essentially a null hypothesis
statistical test based on a Bernoulli model. If the resulting *p*-value is below the desired
significance level, the null hypothesis is rejected and the performance difference between two
algorithms is considered to be statistically significant. By this test, we find that symbolic
KDA statistically significant and outperforms kernel Eigenface and kernel Fisherface methods
at a significance level of .

The receiver operating characteristic (ROC) curve in Figure 2 reports results for a verification scenario. The equal-error rate (EER) is the ROC point at which the false-accept rate is equal to the false-reject rate. The EER for our approach goes from approximately 0.12 to approximately 0.23. The EER for the other methods shows much greater performance degradation.

##### 4.2. Experiments on the Yale Face Database

The experiments were conducted using Yale database to evaluate the excellence of the symbolic KDA for frontal face recognition under different nondark backgrounds. The Yale face database consists of a total of 165 images obtained from 15 different people, with 11 images from each person. Figure 3 shows some typical images of one subject of Yale face database. We preprocessed these images by aligning and scaling them so that the distances between the eyes were the same for all images and also ensuring that the eyes occurred in the same co-ordinates of the image. The resulting image was then cropped. The images were not manually arranged as done in previous set of experiments using ORL database (Section 4.1). In our experiments, 9 images were randomly chosen from each class for training, while the remaining two images were used to construct test symbolic face for each trial.

The experiments were conducted using two different kernels, namely, polynomial kernel and Gaussian kernel. The order of the polynomial kernel should be 2 and the width of Gaussian kernel should be four for proposed symbolic KDA with respect to a minimum distance classifier.

After finding optimal kernel parameter (degree 2) for the symbolic KDA method, the experiments were conducted to find optimal subspace for proposed symbolic KDA, kernel Fisherface, and kernel Eigenface method. The recognition rates, training time, and optimal subspace dimension for Kernel Fisherface, Kernel Eigenface, and symbolic KDA are listed in Table 3. From Table 3, the symbolic KDA method with polynomial degree two has recognition rate 89.00% using only 15 features, where as Kernel Fisherface method requires 42 features to achieve 87.15% recognition rate. This is due to the fact that first few eigenvectors of symbolic PCA account for highest variance of training samples and these few eigenvectors are enough to represent image for recognition purposes. Hence, improved recognition results can be achieved at less computational cost by using symbolic KDA, by virtue of its low dimensionality. The experimental results obtained using Yale face database shows that the proposed symbolic KDA performs well on images with non dark backgrounds.

#### 5. Conclusions

In this paper, we propose a novel symbolic KDA method for face recognition. Symbolic data representation of face images using interval-type features are desirable facial features to cope up with the variations due to illumination, orientation, and facial expression changes. The feasibility of the symbolic KDA has been tested successfully on frontal face images of ORL and Yale databases. Experimental results show that symbolic KDA method with polynomial kernel leads to improved recognition rate at reduced computational cost.

The proposed symbolic KDA has many advantages compared to other popular appearance-based methods. The drawback of other appearance-based methods is that in order to recognize a face seen from a particular pose and under a particular illumination, the face must have been previously seen under the same conditions. The symbolic KDA overcomes this limitation by representing the faces by interval-type features so that even the faces seen previously in different poses, orientations, and illuminations are recognized. Another important merit is that we can use more than one probe image with inherent variability of a face for face recognition, this yields improved recognition rate. This is clearly evident from the experimental results. We observe from the experimental results that the proposed symbolic KDA method yields improved recognition rate in terms of time and feature reduction compare to other kernel-based methods.

The main drawback of our proposed symbolic KDA method is that pose variation is limited up to 20 degree orientation and the performance of proposed method decreases on face images with pose variation greater than 20 degree orientation. The proposed method did not achieve 100% accuracy, this is due to the fact that while constructing the symbolic faces, there may be chance of misalignment of coordinates of eyes, mouth, and nose because of different facial expressions in training images. It can be observed in experimental results obtained using Yale face database, which contains face images with different facial expressions. Moreover, the performance of the proposed symbolic KDA method decreases on images with more variation in facial expressions of Yale face database compared to performance on images with less variation in facial expressions of ORL face database.

#### Acknowledgment

The authors are indebted to the referees for their helpful comments and suggestions, which improved the earlier version of the paper.

#### References

- K.-R. Muller, S. Mika, G. Ratsch, K. Tsuda, and B. Schölkopf, “An introduction to kernel-based learning algorithms,”
*IEEE Transactions on Neural Networks*, vol. 12, no. 2, 181 pages, 2001. View at Publisher · View at Google Scholar - B. Schölkopf, A. Smola, and K.-R. Müller, “Nonlinear component analysis as a kernel eigenvalue problem,”
*Neural Computation*, vol. 10, no. 5, 1299 pages, 1998. View at Publisher · View at Google Scholar - S. Mika, G. Rätsch, J. Weston, B. Schölkopf, and K.-R. Müller, “Fisher discriminant analysis with kernels,” in
*Proceedings of the 9th IEEE Workshop on Neural Networks for Signal Processing (NNSP '99)*, p. 41, Madison, Wis, USA, August 1999. View at Publisher · View at Google Scholar - H.-H. Bock and E. Diday, Eds.,
*Analysis of Symbolic Data*, Springer, Heidelberg, Germany, 2000. - P. S. Hiremath and C. J. Prabhakar, “Symbolic factorial discriminant analysis for face recognition under variable lighting,” in
*Proceedings of the 3rd International Conference on Visual Information Engineering (VIE '06)*, p. 150, IET, Bangalore, India, September 2006. - P. S Hiremath and C. J. Prabhakar, “Acquiring non linear subspace for face recognition using symbolic kernel PCA method,”
*Journal of Symbolic Data Analysis*, vol. 4, no. 1, 15 pages, 2006. View at Google Scholar - P. S. Hiremath, A. Dhanti, and C. J. Prabhakar, “Modeling uncertainty in representation of facial features for face recognition,” in
*Face Recognition*, K. Delac and M. Grgic, Eds., chapter 10, p. 183, ITech, Vienna, Austria, 2007. View at Google Scholar - M.-H. Yang, N. Ahuja, and D. Kriegman, “Face recognition using kernel eigenfaces,” in
*Proceedings of the IEEE International Conference on Image Processing (ICIP '00)*, vol. 1, p. 37, Vancouver, BC, Canada, September 2000. View at Publisher · View at Google Scholar - M.-H. Yang, “Kernel eigenfaces vs. kernel fisherfaces: face recognition using kernel methods,” in
*Proceedings of the 5th IEEE International Conference on Automatic Face and Gesture Recognition (FGR '02)*, p. 215, Washington, DC, USA, May 2002.