Literature DB >> 23936418

Improved minimum squared error algorithm with applications to face recognition.

Qi Zhu1, Zhengming Li, Jinxing Liu, Zizhu Fan, Lei Yu, Yan Chen.   

Abstract

Minimum squared error based classification (MSEC) method establishes a unique classification model for all the test samples. However, this classification model may be not optimal for each test sample. This paper proposes an improved MSEC (IMSEC) method, which is tailored for each test sample. The proposed method first roughly identifies the possible classes of the test sample, and then establishes a minimum squared error (MSE) model based on the training samples from these possible classes of the test sample. We apply our method to face recognition. The experimental results on several datasets show that IMSEC outperforms MSEC and the other state-of-the-art methods in terms of accuracy.

Entities:  

Mesh:

Year:  2013        PMID: 23936418      PMCID: PMC3735590          DOI: 10.1371/journal.pone.0070370

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.240


Introduction

The minimum squared error based classification (MSEC) is sound in theory and is able to achieve a high accuracy [1], [2]. It has been proven that for two-class classification MSEC is identical to linear discriminant analysis (LDA) under the condition that the number of training samples approximates the infinity [1], [2]. In addition, MSEC can be applied to multi-class classification by using a special class label matrix [3]. Various improvements to MSEC such as orthogonal MSEC [4] and kernel MSEC [5]–[8] have been proposed. The MSEC has been applied to a number of problems such as imbalanced classification [7], palm-print verification [9], low-rank representation [10], [11], super-resolution learning [12], image restoration [13], and manifold learning [14]. In recent years, representation based classification (RC) method [15]–[18] has attracted increasing attention in pattern recognition. The main difference between RC and MSEC is that RC tries to use the weighted sum of all the training samples to represent the test sample, whereas MSEC aims to map the training samples to their class labels. RC can be categorized into two types. The first type is the so-called sparse representation method (SRM) such as the methods proposed in [19], [20]. The goal of SRM is to simultaneously minimize the norm of the weight vector and the representation error that is the deviation between constructed sample and test sample. The second type is the so-called non-sparse representation method such as the methods proposed in [21]–[26]. The goal of the non-sparse representation method is to simultaneously minimize the norm of the weight vector and the representation error. The non-sparse representation method has a closed-solution and is usually more computationally efficient than SRM [21]. In this paper, we focus on multi-class classification problem and propose an improved minimum squared error based classification (IMSEC) method. The basic idea of IMSEC is to select a subset of training samples that are similar to the test sample and then build the MSE model based on them. The advantage of the IMSEC is that it seeks the optimal classifier for each test sample. However, MSEC categorizes all the test samples based on a unique classifier. Therefore IMSEC has better performance than MSEC.

The minimum Squared Error Based Classification for Multi-class Problems

Suppose that there are training samples from classes. Let the -dimensional row vector denote the -th training sample, where . We use a -dimensional row vector to represent the class label of the -th training sample. If this sample is from class , the -th entry in is one and the other entries are all zeroes. If a mapping can approximately transform each training sample into its class label, we havewhere , . Clearly, is an matrix, is an matrix, and is a matrix that is to be solved. As Eq. (1) cannot be directly solved, we convert it into the following equation: If is non-singular, can be solved by In general, we use to obtain a stable numerical solution, where and denote a small positive constant and the identity matrix, respectively. Finally, we classify a test sample as follows: the class label of is predicted using , and then the Euclidean distance between and the class label of each class is calculated, respectively. The class label of the-th class is a row vector whose-th element is one and whose other elements are all zeros (). Among the classes, if is closest to the -th class, then is classified into the -th class.

The Algorithm of Improved Minimum Squared Error Based Classification

Suppose the -th class has training samples. Let be the -th training sample of the -th class, where ,. The algorithm of IMSEC has the following three steps.

Step 1

Determine possible classes of the test sample, where . First, the test sample is represented as a weighted sum of the training samples of each class, respectively. For the-th class, it is assumed that is approximately satisfied. can be rewritten as , where and . Then we have . is the representation error between the training samples of -th class and the test sample. The classes that have the smallest representation errors are determined, and they are referred to as base classes.

Step 2

Use the base classes to establish the following MSE modelwhere is composed of all the training samples of the base classes, and is composed of the class labels of these training samples. is computed using . and are a small positive constant and identity matrix, respectively.

Step 3

Exploit and to classify the test sample . The class label of this test sample can be predicted by using . Calculate the Euclidean distance between and the class label of each base class, respectively. Let denote the distance of the between and the class label of the -th class. If , then test sample is assigned into the - th class.

Analysis of the Proposed Method

The proposed method and the MSEC have the following differences. MSEC attempts to obtain a unique model for all the test samples, whereas the proposed method constructs a special MSE for each test sample. MSEC tries to minimize the mean square error between the predicted class labels and the true class labels of the training samples. That means MSEC is capable of mapping the training samples to the correct class labels. However, this does not imply that the model of MSEC can map the test sample to the correct class label accurately. Since the test sample and the training samples that are “close” to the test sample have the similar MSE models, it can be expected that IMSEC performs better in mapping the test sample to the correct class label than CMSE. The proposed method works in the way of coarse-to-fine classification. In detail, step 1 of the proposed method indeed roughly identifies the possible classes of the test sample. Step 2 of the proposed method assigns the test sample into one of the possible classes. For the complicated classification problem, the way of coarse-to-fine classification is usually more effective than the way in one step [27]–[29]. It is worth pointing out that the proposed method is different from CRC [21] and linear regression based classification (LRC) [30]. The proposed method tries to establish a model to map the training samples to their true class labels, whereas CRC uses the weighted combination of all the training samples to represent the test sample, and LRC uses the class-specific training samples to represent the test sample. Moreover, when classifying a test sample, the proposed method and LRC need to solve one and MSE models, respectively, where is the number of the classes. As a result, the proposed method is more efficient than LRC.

Experiments

A. Ethics Statement

Some face datasets were used in this paper to verify the performance of our method. These face datasets are publicly available for face recognition research, and the consent was not needed. The face images and the experimental results are reported in this paper without any commercial purpose.

B. Experimental Results

Face recognition has become a popular pattern classification task. We perform the experiments on ORL, FERET and AR face databases. Our method, CMSE, CRC, SRC, Eigenface [31], Fisherface [32], Nearest Neighbor Classifier (1-NN), 2DPCA [33], Alternative-2DPCA [34], 2DLDA [35], Alternative-2DLDA [36] and 2DPCA+2DLDA [37] were tested in the experiments. Before implementing each method, we converted every face image into a unit vector with the norm of 1. When CRC was implemented, the regular parameter was set to 0.001. In Eigenface method, we used the first 50, 100…, 400 Eigenfaces for feature extraction, respectively, and reported the lowest error rate. In the 2D based subspace methods, including 2DPCA, Alternative-2DPCA, 2DLDA, Alternative-2DLDA and 2DPCA+2DLDA, the number of the projection axes was set to 1,2,…,5, and the lowest error rate was reported. In the ORL database, there are 40 subjects and each subject has 10 different images. For some subjects, the images were taken at different times, varying the lighting, facial expressions (open/closed eyes, smiling/not smiling) and facial details (glasses/no glasses). All the images were taken against a dark homogeneous background with the subjects in an upright, frontal position (with tolerance for some side movement). Each face image contains 92112 pixels, with 256 grey levels per pixel [38]. We resized each face image into a 46 by 56 matrix. Figure 1 shows the face images of one subject in the ORL database. We took the first three, four, five and six face images of each subject as training images and treated the others as test images, respectively. In our method, was set to .
Figure 1

The face images of one subject in the ORL database.

For AR face database, we used 3120 gray face images from 120 subjects, each providing 26 images [39]. These images were taken in two sessions and show faces with different facial expressions, in varying lighting conditions and occluded in several ways. Figure 2 shows the 26 face images of one subject in the AR database. We took the first four, five, six, seven and eight face images of each subject as training images and treated the others as test images, respectively. In our method, was set to .
Figure 2

The face images of one subject in the AR database.

A subset of the FERET face database is used to test our method. This subset includes 200 subjects, and each subject has 7 images. It is composed of the images whose names are marked with two-character strings: ‘ba’, ‘bj’, ‘bk’, ‘be’, ‘bf’, ‘bd’, and ‘bg’. This subset involves variations in facial expression, illumination, and pose [40]. The facial portion of each original image was cropped to form a 4040 image. Figure 3 shows some face images from the FERET database. We took the first five and six face images of each subject as training images and treated the others as test images, respectively. In our method, was set to .
Figure 3

Some face images from the FERET database.

Tables 1, 2 and 3 show the classification error rates of the methods on the ORL, AR and FERET databases, respectively. We can observe that our method always obtains the lowest classification error rate. In other words, our method can achieve the desirable classification result.
Table 1

Rates of classification errors of the methods on the ORL database (%).

Number of the originaltraining samplesper class3456
The proposed method11.075.834.501.87
CMSE13.937.927.503.75
CRC15.369.178.005.63
SRC19.2915.0014.5011.87
Eigenface26.0720.0014.0010.00
Fisherface23.0122.6423.299.08
1-NN20.3615.0014.008.75
2DPCA14.2911.259.503.75
Alternative-2DPCA13.9310.428.503.75
2DLDA11.797.929.504.37
Alternative-2DLDA17.5013.7513.504.37
2DPCA+2DLDA16.0712.5010.004.37
Table 2

Rates of classification errors of the methods on the AR database (%).

Number of the original training samples per class4567
The proposed method25.2723.2922.9620.92
CMSE27.9224.8825.8725.48
CRC29.8928.0229.7128.90
SRC41.9743.4134.0429.78
Eigenface41.7847.6624.7926.05
Fisherface44.1740.7125.1323.89
1-NN37.6939.4025.8725.04
2DPCA40.3841.8730.8331.36
Alternative-2DPCA40.2340.8730.6331.54
2DLDA50.6852.2235.3333.25
Alternative-2DLDA54.0955.8341.9636.40
2DPCA+2DLDA35.5337.9026.4228.03
Table 3

Rates of classification errors of the methods on the FERET database (%).

Number of the originaltraining samplesper class56
The proposed method19.003.50
CMSE29.257.50
CRC38.7529.50
Eigenface37.0036.00
Fisherface47.5061.00
SRC59.2552.50
1-NN28.5027.00
2DPCA35.2549.50
Alternative-2DPCA36.0050.00
2DLDA29.2525.25
Alternative-2DLDA29.2531.50
2DPCA+2DLDA30.5035.00

Conclusions

The proposed method, i.e. IMSEC, establishes a special MSE model for each test sample. When building the classification model, IMSEC uses only the training samples that are close to the test sample. Theoretical analyses were presented to explore the properties of IMSEC. Compared with MSEC that classifies all the test samples based on a unique model, IMSEC can perform better in classifying the test samples. We tested the proposed method on three face datasets. The experimental results clearly demonstrated that IMSEC outperforms MSEC and the other state-of-the-art methods.
  9 in total

1.  Two-dimensional PCA: a new approach to appearance-based face representation and recognition.

Authors:  Jian Yang; David Zhang; Alejandro F Frangi; Jing-yu Yang
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2004-01       Impact factor: 6.226

2.  Linear regression for face recognition.

Authors:  Imran Naseem; Roberto Togneri; Mohammed Bennamoun
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2010-11       Impact factor: 6.226

3.  A coarse-to-fine strategy for multiclass shape detection.

Authors:  Yali Amit; Donald Geman; Xiaodong Fan
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2004-12       Impact factor: 6.226

4.  Sparse representation for coarse and fine object recognition.

Authors:  Thang V Pham; Arnold W M Smeulders
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2006-04       Impact factor: 6.226

5.  Eigenfaces for recognition.

Authors:  M Turk; A Pentland
Journal:  J Cogn Neurosci       Date:  1991       Impact factor: 3.225

6.  Toward a practical face recognition system: robust alignment and illumination by sparse representation.

Authors:  Andrew Wagner; John Wright; Arvind Ganesh; Zihan Zhou; Hossein Mobahi; Yi Ma
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2012-02       Impact factor: 6.226

7.  Sparse approximation to the eigensubspace for discrimination.

Authors:  Zhihui Lai; Wai Keung Wong; Zhong Jin; Jian Yang; Yong Xu
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2012-12       Impact factor: 10.451

8.  A generalized accelerated proximal gradient approach for total-variation-based image restoration.

Authors:  Wangmeng Zuo; Zhouchen Lin
Journal:  IEEE Trans Image Process       Date:  2011-03-24       Impact factor: 10.856

9.  Robust face recognition via sparse representation.

Authors:  John Wright; Allen Y Yang; Arvind Ganesh; S Shankar Sastry; Yi Ma
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2009-02       Impact factor: 6.226

  9 in total
  2 in total

1.  Constructive autoassociative neural network for facial recognition.

Authors:  Bruno J T Fernandes; George D C Cavalcanti; Tsang I Ren
Journal:  PLoS One       Date:  2014-12-26       Impact factor: 3.240

2.  Blurred palmprint recognition based on stable-feature extraction using a Vese-Osher decomposition model.

Authors:  Danfeng Hong; Jian Su; Qinggen Hong; Zhenkuan Pan; Guodong Wang
Journal:  PLoS One       Date:  2014-07-03       Impact factor: 3.240

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.