| Literature DB >> 29293581 |
Yan Yan1, Feifei Lee1, Xueqian Wu1, Qiu Chen2.
Abstract
In this paper, we propose a face recognition algorithm based on a combination of vector quantization (VQ) and Markov stationary features (MSF). The VQ algorithm has been shown to be an effective method for generating features; it extracts a codevector histogram as a facial feature representation for face recognition. Still, the VQ histogram features are unable to convey spatial structural information, which to some extent limits their usefulness in discrimination. To alleviate this limitation of VQ histograms, we utilize Markov stationary features (MSF) to extend the VQ histogram-based features so as to add spatial structural information. We demonstrate the effectiveness of our proposed algorithm by achieving recognition results superior to those of several state-of-the-art methods on publicly available face databases.Entities:
Mesh:
Year: 2018 PMID: 29293581 PMCID: PMC5749794 DOI: 10.1371/journal.pone.0190378
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Approaches mentioned in the Introduction.
| Approach | Advantages | Limitations | |
|---|---|---|---|
| Acronyms | Algorithms | ||
| Ref. | |||
| PCA | Eigenfaces | •Low computation time. | •Contains no class information on the input data. |
| [ | •Fails to capture high-order statistics. | ||
| LDA | Fisherfaces | •Includes class-specific discriminatory information. | •Suffers from the small sample size problem. |
| [ | •Fails to capture high-order statistics. | ||
| 2DPCA | Two-dimensional PCA | •Can directly extract the matrix features of 2D images. | •More coefficients than PCA for image representation. |
| •Lower dimensionality than PCA. | |||
| [ | •More computationally efficient than PCA. | ||
| 2DLDA | Two-dimensional LDA | •Can directly extract the matrix features of 2D images. | •More coefficients than LDA for image representation. |
| [ | •More computationally efficient and stable than LDA. | ||
| NMF | Non-negative Matrix Factorization | •Can capture important local differences. | •Sensitivity to facial variations. |
| [ | |||
| LPP | Locality Preserving Projections | •Can find the intrinsic low-dimensional nonlinear manifold structure hidden in the observation space. | •Sensitivity to facial variations. |
| [ | |||
| SPP | Structure-Preserved Projections | •Can preserve the configural structure of facial image in subspace. | •Robust to variations such as head, pose, lighting condition, and facial expression. |
| [ | |||
| LBP | Local Binary Patterns | •Simple calculation. | •Sensitivity to noise. |
| •Good for extracting the local texture features of a face image. | •Features contain no shape information. | ||
| [ | •Invariant to rotation and grey-scale. | ||
| HOG | Histograms of Oriented Gradients | •Invariant to illumination and 2D rotation. | •Non-robust to scale changes. |
| [ | |||
| EBGM | Elastic Bunch Graph Matching | •Can model a face as a 2-D elastic graph. | •Non-robust to changes in expression and illumination. |
| [ | •High computation cost. | ||
| SIFT | Scale Invariant Feature Transform | •Robust to rotation and scale changes. | •High computation cost. |
| [ | |||
| DCT | Discrete Cosine Transform | •Data-independent. | •Complex calculation. |
| [ | •Can be implemented using a fast algorithm. | ||
| DCV | Discriminative Common Vectors | •Efficiency (real-time). | •Applications for under-sampled data are limited. |
| •Numerical stability. | •Linear technique (Inadequate to describe the complexity of face image due to facial variations). | ||
| [ | •Can handle the small sample size problem. | ||
| SRC | Sparse Representation Classification | •Can correct corruptions possibly existing in testing data. | •Cannot handle cases in which the training data are corrupted. |
| [ | • Computationally expensive. | ||
| LRC | Linear Regression Classification | •Simple architecture. | •Non-robust to severe illumination. |
| [ | •Computationally efficient. | ||
| VPC | Vector Projection Classification | •Simple architecture. | •Non-robust to severe illumination. |
| [ | •Computationally efficient. | ||
| NDC | Nearest Distance Classifier | •Computationally efficient. | •Lazy learning. |
| [ | |||
| BC | Bayesian Classifier | •Simple calculation. | •Need prior probabilities. |
| •Few estimated parameters. | |||
| [ | •Insensitive to missing data. | ||
| SVM | Support Vector Machines | •Efficient in classification with nonlinear data. | •Low efficiency in handling large-scale training samples. |
| [ | •Low efficiency in solving multi classification problems. | ||
| CNN | Convolution Neural Network | •Robust to rotation, translation and scaling deformation of images. | •Require a large number of training samples. |
| [ | •Hardware requirements. | ||
Fig 1Face recognition process using the VQ algorithm.
Fig 2Face recognition process using the MSR-MSF-VQ algorithm.
Fig 3Face image partition strategies based on several equal-sized sub-regions.
Face recognition accuracy based on seven cases.
| Approach | Recognition rate (%) |
|---|---|
| VQ | 95.600 |
| MSF-VQ(0) | 95.435 |
| MSF-VQ(45) | 95.442 |
| MSF-VQ(90) | 96.278 |
| MSF-VQ(135) | 96.254 |
| MSF-VQ(mix) | 95.722 |
| MSF-VQ(ave) | 96.153 |
Fig 4The average recognition rate using different values of d.
Fig 5The average recognition rate using different values of n.
Fig 6(a) Face recognition rate achieved by varying d and fixing n = 50 on the FERET database. (b) Face recognition rate achieved by varying n and fixing d = 1 on the FERET database.
Influence of different factors on the face recognition rate using the MSF-VQ algorithm.
| Image size | |||||||||||
| Zoom Factor | 0.5 | 0.6 | 0.7 | 0.8 | 0.9 | 1 | 1.1 | 1.2 | 1.3 | 1.4 | 1.5 |
| Rank-one recognition rate (%) | 82.8 | 83.9 | 84.4 | 84.8 | 85.2 | 86.1 | 86.5 | 86.6 | 86.9 | 86.0 | 86.0 |
| Similarity measures | |||||||||||
| Measure Method | Euclidean | Manhattan | Formula in [ | ||||||||
| Rank-one recognition rate (%) | 84.6 | 86.1 | 84.5 | ||||||||
| Directions of the occurrence matrix | |||||||||||
| Direction | 0 | 45 | 90 | 135 | mix | ave | |||||
| Rank-one recognition rate (%) | 85.2 | 85.8 | 86.7 | 85.5 | 86.1 | 85.9 | |||||
Recognition results using different segmentation strategies on the FERET database (results obtained with our proposed algorithm are in bold).
| Division strategies | Image size | Rank-one Recognition rate (%) |
|---|---|---|
| 1×1 | 204×204 | 86.3 |
| 2×2 | 204×204 | 95.3 |
| 3×3 | 204×204 | 97.3 |
| 4×4 | 204×204 | 97.5 |
| 6×6 | 204×204 | 98.1 |
| 7×7 | 203×203 | 97.8 |
Performance comparison on the FERET face database (results obtained with our proposed algorithm are in bold).
| Approach | Rank-one Recognition rate (%) |
|---|---|
| LDA | 72.1 |
| Bayesian MAP | 81.7 |
| Bayesian ML | 81.7 |
| PCA Mahalanobis Cosine | 85.3 |
| Gabor-EBGM | 87.3 |
| HOG | 90.0 |
| LBP | 93.0 |
| PCA Euclidean | 94.3 |
| HOG-EBGM | 95.5 |
| SIFT | 95.9 |
Recognition rates of different approaches on the FERET database.
| Methods | Feature dimension | Recognition Accuracy (%) |
|---|---|---|
| AlexNet+Manhattan | 4096 | 89.5 |
| AlexNet+SVM | 4096 | 93.97 |
| CenterlossNet+Manhattan | 512 | 97.0 |
| CenterlossNet+SVM | 512 | 99.0 |
| MSR-MSF-VQ+Manhattan | 1650 | 98.2 |
| MSR-MSF-VQ+SVM | 1650 | 99.16 |
Recognition times of different approaches.
| Method | Feature extraction time (ms) | Total time (ms) |
|---|---|---|
| MSF-VQ | 381 | 1883 |
| AlexNet | 1360 | 47795 |
| CenterlossNet | 1917 | 9587 |
The recognition rates of different approaches on the AR database.
| Approach | Sunglasses (%) | Scarf (%) |
|---|---|---|
| DCV | 13.33 | 10 |
| Fisherfaces | 27.33 | 23.67 |
| DCT | 29.67 | 5.33 |
| Eigenfaces | 42.33 | 7.33 |
| SRC | 55.67 | 27.67 |
| LRC | 59 | 9.33 |
| VPC | 62.67 | 6.67 |
The recognition rates of different approaches and different partitioning strategies.
| Approach | Sunglasses (%) | Scarf (%) |
|---|---|---|
| MLRC | 67.33 | 95.33 |
| LVPC | 70.00 | 83.33 |
| MSR-MSF-VQ-4 | 89.00 | 29.70 |
| MSR-MSF-VQ-16 | 100 | 94.30 |
| MSR-MSF-VQ-25 | 99.7 | 98.0 |
Performance comparison on the Yale face database (results of our proposed algorithm are in bold).
| Approach | Rank-one Recognition rate (%) |
|---|---|
| LPP | 75.00 |
| S-LPP | 75.73 |
| NMF | 77.27 |
| SubXPCA | 77.80 |
| PCA | 78.00 |
| ModPCA | 83.47 |
| SpPCA | 83.87 |
| Aw-SpPCA | 84.93 |
| SpSLPP | 86.13 |
| SpNMF | 88.40 |
| SPP | 93.33 |
Algorithms compared in our experiments on the Yale database.
| No. | Algorithms | Acronyms |
|---|---|---|
| 1 | Principle Component Analysis | PCA [ |
| 2 | Modular Principle Component Analysis | ModPCA [ |
| 3 | Sub-pattern Principle Component Analysis | SpPCA [ |
| 4 | Adaptively Weighted Sub-pattern Principle Component Analysis | Aw-SpPCA [ |
| 5 | Cross-sub-pattern correlation based Principle Component Analysis | SubXPCA [ |
| 6 | Non-negative Matrix Factorization | NMF [ |
| 7 | Sub-pattern Non-negative Matrix Factorization | SpNMF [ |
| 8 | Sub-pattern based Spatially Smooth Locality Preserving Projections | SpSLpp [ |
| 9 | Locality Preserving Projections | LPP [ |
| 10 | Spatially Smooth Locality Preserving Projections | S-LPP [ |
| 11 | structure-preserved projections | SPP [ |
Fig 7Mean recognition accuracy comparison on the Yale face database.
Performance comparison of the first experiment (results of our proposed algorithm are in bold).
| Approach | Subset 2 (%) | Subset 3 (%) | Subset 4 (%) | Subset 5 (%) |
|---|---|---|---|---|
| Raw image | 95.83 | 76.67 | 46.67 | 24.24 |
| HEQ | 100 | 97.5 | 75 | 60 |
| wavelet-based normalization | 100 | 100 | 94.76 | 90.83 |
| MSR-MSF-VQ | 100 | 97.5 | 66.4 | 21.6 |
Performance comparison of the second experiment (results of our proposed algorithm are in bold).
| Approach | Subset 1 (%) | Subset 2 (%) | Subset 3 (%) | Subset 4 (%) | Subset 5 (%) |
|---|---|---|---|---|---|
| ORI | 98.6 | 93.3 | 43.3 | 17.9 | 10.5 |
| HE | 100 | 99.2 | 73.3 | 42.1 | 43.2 |
| RG | 100 | 100 | 94.2 | 59.3 | 39.5 |
| LTV | 100 | 100 | 75.8 | 72.1 | 79.8 |
| GradFace | 100 | 100 | 99.2 | 94.3 | 98.9 |
| MSR-MSF-VQ | 100 | 100 | 83.3 | 54.3 | 17.4 |
The maximal accuracy recognition rates (%) on the CAS-PEAL-R1 face database.
| Approach | Accessory | Expression | Lighting |
|---|---|---|---|
| TVQI | 43.41 | 60.06 | 10.97 |
| TV_L+HE | 46.48 | 61.27 | 7.40 |
| TV_L+HE | 47.27 | 61.02 | 7.31 |
| TV_L+RHE | 48.23 | 58.34 | 8.69 |
| MSR-MSF-VQ-Manhattan | 66.6 | 90.8 | 6.7 |