| Literature DB >> 24634613 |
Abstract
Maximum margin criterion (MMC) is a well-known method for feature extraction and dimensionality reduction. However, MMC is based on vector data and fails to exploit local characteristics of image data. In this paper, we propose a two-dimensional generalized framework based on a block-wise approach for MMC, to deal with matrix representation data, that is, images. The proposed method, namely, block-wise two-dimensional maximum margin criterion (B2D-MMC), aims to find local subspace projections using unilateral matrix multiplication in each block set, such that in the subspace a block is close to those belonging to the same class but far from those belonging to different classes. B2D-MMC avoids iterations and alternations as in current bilateral projection based two-dimensional feature extraction techniques by seeking a closed form solution of one-side projection matrix for each block set. Theoretical analysis and experiments on benchmark face databases illustrate that the proposed method is effective and efficient.Entities:
Mesh:
Year: 2014 PMID: 24634613 PMCID: PMC3920850 DOI: 10.1155/2014/875090
Source DB: PubMed Journal: ScientificWorldJournal ISSN: 1537-744X
Figure 1Images and their blocks of the first subject from ORL database: (a) 10 images of size 112 × 92; (b)–(e) 4 image blocks of size 28 × 92 for each image in (a).
Algorithm 1
Figure 2Images of one person from the ORL face database.
Figure 3Images of one person from the Yale face database.
Face recognition accuracies of different methods on the ORL database. TN means number of training samples per subject, and the number in brackets is the corresponding projection dimensionality. The bold value means the highest accuracy among all the methods.
| Method | TN = 2 | TN = 3 | TN = 4 |
|---|---|---|---|
| PCA | 70.67% (79) | 78.88% (118) | 84.21% (152) |
| LDA | 72.80% (25) | 83.79% (39) | 90.13% (39) |
| MMC | 77.97% (39) | 86.32% (39) | 91.63% (39) |
| GLRAM | 71.30% (17 × 17) | 79.84% (11 × 11) | 84.73% (16 × 16) |
| 2DLDA | 78.13% (11 × 11) | 86.79% (16 × 16) | 92.08% (15 × 15) |
| 2DMMC | 78.75% (12 × 12) | 87.50% (10 × 10) |
|
| B2D-MMC |
|
| 92.88% ( |
Face recognition accuracies of different methods on the Yale database. TN means number of training samples per subject, and the number in brackets is the corresponding projection dimensionality. The bold value means the highest accuracy among all the methods.
| Method | TN = 2 | TN = 3 | TN = 4 |
|---|---|---|---|
| PCA | 46.04% (29) | 49.96% (44) | 55.67% (58) |
| LDA | 42.81% (11) | 60.33% (14) | 68.10% (13) |
| MMC | 52.37% (14) | 61.83% (14) | 67.95% (15) |
| GLRAM | 49.33% (6 × 6) | 54.17% (6 × 6) | 57.76% (5 × 5) |
| 2DLDA | 44.37% (7 × 7) | 59.71% (5 × 5) | 68.71% (5 × 5) |
| 2DMMC | 54.37% (6 × 6) | 63.50% (9 × 9) | 68.86% (15 × 15) |
| B2D-MMC |
|
|
|
Figure 4Training time of two methods on ORL and Yale datasets with TN = 2.
Figure 5Recognition performance on the ORL dataset for different number of blocks per image.
Figure 6Recognition performance on the Yale dataset for different number of blocks per image.