| Literature DB >> 23755136 |
Junfeng Gao1, Zhao Wang, Yong Yang, Wenjia Zhang, Chunyi Tao, Jinan Guan, Nini Rao.
Abstract
A new machine learning method referred to as F-score_ELM was proposed to classify the lying and truth-telling using the electroencephalogram (EEG) signals from 28 guilty and innocent subjects. Thirty-one features were extracted from the probe responses from these subjects. Then, a recently-developed classifier called extreme learning machine (ELM) was combined with F-score, a simple but effective feature selection method, to jointly optimize the number of the hidden nodes of ELM and the feature subset by a grid-searching training procedure. The method was compared to two classification models combining principal component analysis with back-propagation network and support vector machine classifiers. We thoroughly assessed the performance of these classification models including the training and testing time, sensitivity and specificity from the training and testing sets, as well as network size. The experimental results showed that the number of the hidden nodes can be effectively optimized by the proposed method. Also, F-score_ELM obtained the best classification accuracy and required the shortest training and testing time.Entities:
Mesh:
Year: 2013 PMID: 23755136 PMCID: PMC3670874 DOI: 10.1371/journal.pone.0064704
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Figure 1The preprocessing results of a guilty subject and an innocent subject.
1A: Three averaged waves over the three kinds of stimuli respectively at Pz electrode from the guilty subject. 1B: The averaged waves over the three kinds of stimuli respectively at Pz electrode from the innocent subject. 1C: The brain topographies at the latency of 348 ms of the averaged P responses (the solid line in Figure 1A). 1D: The brain topographies at the peak point of 316 ms of the averaged P responses (the solid line in Figure 1B).
Figure 2SLFN with K hidden, n input and m output nodes.
Figure 3The block diagram of the proposed method F-score_ELM.
The results of feature valuation on original 31 features using F-score method.
| Features |
|
|
|
|
|
| 0.581 |
|
| 0.162 |
|
| 0.019 |
|
|
|
|
| 0.124 |
|
| 0.001 |
|
| 0.305 |
|
|
|
|
| 0.075, 0.001, 0.317, 0.002, 0.075, |
|
| 0.003, 0.154, 0.097, 0.069, 0.364, |
|
|
|
|
|
|
Performance of the classification models with the optimal NFS and NHN (or NSV).
| Classification models |
|
|
|
| ||||
|
|
| Training | Testing | |||||
|
|
|
|
| |||||
|
| 1.25 | 0.51 | 95.31±1.38 | 88.61±4.44 | 95.05±3.25 | 86.38±3.15 | 45 | 31 |
|
| 20.22 | 22.06 | 97.88±0.41 | 97.68±0.41 | 96.33±1.87 | 95.15±2.29 | 58.28 | 31 |
|
| 0.03 | 0.004 | 98.72±0.35 | 98.16±0.51 | 98.26±0.32 | 98.14±0.44 | 51 | 31 |
|
| 106.64 | 1.59 | 95.10±1.82 | 95.75±1.97 | 98.20±1.82 | 98.34±1.93 | 20 | 14 |
|
| 675.58 | 25.23 | 99.32±0.35 | 99.28±0.33 | 99.65±0.02 | 98.91±0.02 | 57.76 | 14 |
|
| 0.64 | 0.014 | 98.82±0.36 | 99.31±0.36 | 98.99±0.48 | 98.79±0.51 | 30 | 13 |
|
| 58.58 | 0.184 | 95.26±0.23 | 88.05±1.36 | 95.80±3.92 | 90.05±5.27 | 34 | 25 |
|
| 633.77 | 21.62 | 98.06±0.48 | 98.21±0.44 | 97.68±0.02 | 97.95±0.02 | 60.25 | 25 |
|
| 0.61 | 0.003 | 99.42±0.38 | 98.52±0.72 | 99.27±0.24 | 98.17±0.25 | 29 | 11 |
Balanced accuracy of each model and statistical analysis results between F-score_ELM and the other models.
| Classification models |
| |||
| Training | Testing | |||
|
|
|
|
| |
|
| 91.96 | 0.000* | 90.72 | 0.000* |
|
| 97.78 | 0.000* | 95.74 | 0.000* |
|
| 98.44 | 0.044▴ | 98.20 | 0.008* |
|
| 95.43 | 0.000* | 98.27 | 0.009* |
|
| 99.30 | 0.006* | 99.30 | 0.004* |
|
| 99.06 | 0.233 | 98.89 | 0.173 |
|
| 91.66 | 0.000* | 92.93 | 0.000* |
|
| 98.14. | 0.0008* | 97.82 | 0.006* |
|
| 98.97 | [] | 98.72 | [] |
“*” and “▴” denotes p value<0.01 and p value<0.05, respectively for the comparison of the F-score and the indicated model using paired t-test.
Figure 4Training accuracy and NHN/NSV as a function of NFS achieved by the three classification models.
Each point in each curve corresponds to the highest classification performance of the indicated model with the optimal NHN/NSV. 4A: Highest sensitivity with the optimal NHN or NSV vs NFS. 4B: Highest specificity with the optimal NHN or NSV vs NFS. 4C: NHN vs NFS for which BA_train achieves its highest value. 4D: NSV vs NFS for which BA_train achieves its highest value.
Figure 5Training accuracy (constant NFS = 11) as a function of NHN achieve by the F-score_ELM.
5A: Highest sensitivity vs log (NHN). 5B: Highest specificity vs log (NHN).