Literature DB >> 32256550

Double-Criteria Active Learning for Multiclass Brain-Computer Interfaces.

Qingshan She1, Kang Chen1, Zhizeng Luo1, Thinh Nguyen2, Thomas Potter2, Yingchun Zhang2.   

Abstract

Recent technological advances have enabled researchers to collect large amounts of electroencephalography (EEG) signals in labeled and unlabeled datasets. It is expensive and time consuming to collect labeled EEG data for use in brain-computer interface (BCI) systems, however. In this paper, a novel active learning method is proposed to minimize the amount of labeled, subject-specific EEG data required for effective classifier training, by combining measures of uncertainty and representativeness within an extreme learning machine (ELM). Following this approach, an ELM classifier was first used to select a relatively large batch of unlabeled examples, whose uncertainty was measured through the best-versus-second-best (BvSB) strategy. The diversity of each sample was then measured between the limited labeled training data and previously selected unlabeled samples, and similarity is measured among the previously selected samples. Finally, a tradeoff parameter is introduced to control the balance between informative and representative samples, and these samples are then used to construct a powerful ELM classifier. Extensive experiments were conducted using benchmark and multiclass motor imagery EEG datasets to evaluate the efficacy of the proposed method. Experimental results show that the performance of the new algorithm exceeds or matches those of several state-of-the-art active learning algorithms. It is thereby shown that the proposed method improves classifier performance and reduces the need for training samples in BCI applications.
Copyright © 2020 Qingshan She et al.

Entities:  

Mesh:

Year:  2020        PMID: 32256550      PMCID: PMC7091553          DOI: 10.1155/2020/3287589

Source DB:  PubMed          Journal:  Comput Intell Neurosci


1. Introduction

Brain-computer interfaces (BCIs) are systems that allow users to control external devices via observed brain activity, without relying on peripheral nerve or muscle activity [1]. The most common and useful BCIs are constructed using noninvasive brain activity recording techniques, such as electroencephalography (EEG) [2]. While EEG has become widely used for medical monitoring, rehabilitation, neuroprosthesis, and other healthcare applications [3-5], the data acquisition process can be lengthy and exhaustive for users [6]. In addition, EEG signals often vary over the course of an experiment due to both biological and technical causes, including subject-specific anatomical differences, intersession variability, and the attentional drift of subjects [7]. Consequently, users must often undergo a long data collection process to train a suitable BCI system. This poses a prohibitive burden for individuals with paralysis or a severely injured central nervous system, making it a major hurdle for therapeutic applications. It is therefore of the utmost importance that developed BCI systems achieve efficient and robust performance with as few samples as possible. One approach that has been effectively applied to cases with limited training sets is the introduction of active learning (AL) to the BCI calibration procedure. AL queries the class labels of informative samples within the unlabeled sample space to maximize the efficiency of the learning model, and its application greatly reduces the complexity of training samples without any obvious loss of classification accuracy [8]. In essence, AL is an iterative sampling and labeling procedure. On each iteration, AL extracts the sample or batch of samples that are most valuable for improving the current classification model from the unlabeled data pool, and these samples are then manually labeled. The greatest challenge for AL methods is identifying the most informative samples so that the maximum prediction accuracy can be achieved. A number of sample-selection criteria have then been applied to this task, including (1) query-by-committee (QBC), in which several distinct classifiers are used and the selected samples are those with the largest difference between the labels predicted by different classifiers [9-11]; (2) margin uncertainty sampling, wherein the samples are selected according to the maximum uncertainty based on their respective distances from the classification boundaries [12, 13]; (3) max-entropy sampling, which uses entropy as the uncertainty measure via probabilistic modeling [14, 15]; and (4) diversity sampling, which prefers selecting representative samples [16]. Over the past few decades, many supervised learning models have been adopted as baseline classifiers for AL, including linear discriminate analysis (LDA) [12, 17], support vector machine (SVM) [18, 19], artificial neural network (ANN) [20], and extreme learning machine (ELM) [21, 22]. Among these, the ELM has shown a high learning speed and good generalizability in preliminary testing. Additionally, it can be directly applied to both two-class and multiclass classification. To date, few studies have attempted to introduce AL algorithms into the ELM framework, although these have shown the method to be competitive with active SVMs [13, 14, 23]. Specifically, Yu et al. [13] proposed an active learning method called AL-ELM with the goal of saving training time, and results showed a classification performance comparable to that of AL-SVM [18]. Zhang and Er [23] then introduced the SEAL-ELM method by combining the online sequential ELM (OS-ELM) with AL, yielding a higher classification accuracy than offline combinations of AL and SVM on most test datasets. Regrettably, these existing active ELMs only consider a single-querying strategy, leaving space for improvement. The intuitive next step was to then introduce multiple querying strategies to select desirable samples. In fact, researchers have tried to combine two strategies in AL with base classifier SVM, with each performing better than their single-query counterparts [24-26]. At present, however, few implementations of active learning with ELM have been explored and applied for motor imagery- (MI-) based BCI systems [8, 13]. The present investigation intends to fill this gap by combining a two-query AL algorithm with an ELM and testing the method in a BCI application. A well-defined, general framework for active learning is thereby developed in a manner that accounts for both informativeness and representativeness in a multiclass situation. First, an uncertainty sampling strategy is adopted to select a relative large number of samples using the base ELM classifier. The degree of diversity between labeled training data and previously selected, unlabeled samples is then assessed, along with the degree of similarity between the unlabeled samples. Finally, highly informative and representative samples are used to update the ELM classifier through the introduction of a tradeoff parameter. The method is then tested on several benchmark datasets, along with a multiclass MI EEG dataset from BCI Competition IV Dataset 2a. Results demonstrate that the performance of the new method compares favorably with that of existing AL approaches. Compared to existing ELM-based active learning algorithms, the new method has several noteworthy aspects: Considering that the use of a single uncertainty strategy may not take full advantage of the abundant information with unlabeled data, the AL-ELM algorithm is extended to combine two querying strategies (uncertainty and diversity) in order to select the most valuable samples from the unlabeled EEG data pool. The proposed algorithm provides a straightforward and meaningful way to measure representativeness by assaying two kinds of similarity: the similarity between a query sample and the labeled dataset, and the similarity between any two possible query samples. Employing this modified diversity strategy can help isolate highly representative samples during the active learning process.

2. Background Knowledge

2.1. Active Learning

Active learning methods typically comprise five basic components: L, U, T, Q, and S. L is the limited labeled dataset, U is the pool of samples/instances that contains abundant unlabeled instances, T is the classification model trained by L, Q is a query strategy to select the most valuable instances from U, and S is a human annotator that labels the selected instances correctly. AL is an iterative procedure that gradually adds the most important samples, queried by Q and labeled by S, from U to L to update the classification model T. The iterative AL process will continue in this manner until a predefined criterion is met. The ability to identify both an excellent classification model T and an effective query strategy Q is highly important for active learning algorithms. Depending on the number of querying samples at each iteration, AL can be divided into two groups: stream-based AL and pool-based AL. In stream-based AL, the learner can only access one sample per iteration, while pool-based AL allows the learner to select a batch of samples during each iteration. Adjusting the selection method and number of queried samples then creates different AL algorithms, such as the QBC strategy, the uncertainty strategy, and the diversity strategy.

2.2. Basic ELM

Single-hidden-layer feedforward neural networks (SLFNs) are capable of universal approximation [21]. Consider a dataset containing N training samples, {X, Y}={x , y } , with the input x =[x , x ,…,x ] ∈ ℝ and a corresponding desired output of y =[y , y ,…,y ] ∈ ℝ , where p and q represent the respective dimensions and T denotes a transpose operation. Assuming that M is the number of hidden neurons, the output function of the SLFNs is mathematically modeled aswhere β =[β, β,…,β]T ∈ ℝ is the weight vector that connects the i-th hidden neuron to the output neurons, a =[a , a ,…,a ]T ∈ ℝ is a randomly chosen input weight vector connecting the i-th hidden neuron to the input neurons, b ∈ ℝ(i=1,…, M) is a randomly chosen bias of the i-th hidden node, and g(•) is the activation function, which can be any nonlinear piecewise continuous function (such as a sigmoid function or Gaussian function). For convenience, equation (1) can be rewritten in matrix notation aswhere Y=[y 1, y 2,…,y ]T ∈ ℝ is the expected network output, β=[β 1, β 2,…,β ] ∈ ℝ denotes the weight of output layer, and H is the hidden layer output matrix which is defined as Unlike SLFNs, which require that the parameters of hidden neurons are adjusted during training, ELM adopts randomly generated hidden layer parameters and a tuning-free training strategy [22]. Even with these random hidden node parameters, ELM maintains the universal approximation capability of SLFNs [21]. The ELM training then aims to find suitable network parameters to minimize the approximation error ‖H β − Y‖2. To achieve better generalization performance, a regularization parameter c is introduced in [27], with its corresponding objective function given aswhere ‖·‖2 denotes the l 2-norm of a matrix or a vector. We can obtain the output weight vector β using the Moore-Penrose principle. The solution of equation (4) is then β=((I/c)+H T H)−1 H T Y if N > M, and β=H T((I/c)+H H T)−1 Y if N < M.

3. The D-AL-ELM Method

In this section, we present a novel active learning algorithm, D-AL-ELM, that incorporates both the uncertainty and diversity strategies into consecutive steps. This identifies the most valuable, informative instances, which can then be selected to update the baseline classifier ELM during each learning round.

3.1. Discriminative Information by the Uncertainty Criterion

The uncertainty criterion is used to measure the informativeness of each sample. Uncertain samples which lie along the boundaries of different classes carry more information and play a more significant role in the construction of a classifier. In this implementation, the best-versus-second-best (BvSB) strategy is adopted to estimate the uncertainty of each sample. The BvSB strategy is based on a calculation of posterior probability, which considers the difference in probability values between the two classes with the highest estimated probabilities [28]. The outputs of the ELM then approximate the posterior probabilities of the different classes [13]. To do this, a sigmoid function is used to construct a mapping relationship between the real outputs of the ELM and the posterior probabilities, which is described aswhere f (x) denotes the actual output of the i − th output node corresponding to the time instance x. In practice, equation (5) is only applied to two-class problems, such that the sum of the converted posterior probabilities for the instance x is always 1. However, application in multiclass problem may create a summed posterior proximity that exceeds 1, so calculated probabilities were normalized using the following formula:where p(y=1|f (x)) is the original probability of the i − th class. Based on the above parameters, the BvSB strategy for each sample x can be expressed aswhere p(y best | x) and p(y second−best | x) are the largest and second largest posterior probabilities of x, respectively. It should be noted that f(x) values are inversely related to the amount of uncertainty in a sample, with smaller values indicating greater uncertainty.

3.2. Representative Information by the Diversity Criterion

The selection of redundant or overly similar samples is of little use when attempting to construct a robust classifier. It is therefore necessary to use a diversity criterion to select a batch of samples which are diverse in nature. A feasible way of measuring the diversity of uncertain samples is the cosine angle distance. Following this approach, the similarity between two samples x and x is given by As can be seen from equation (8), the similarity S(x , x ) between the two samples x and x is small if these two samples are far from each other, and vice versa. Suppose a batch of samples W={w 1, w 2,…, w }. If the value of max S(x, w ) is small, then the new sample x is diverse from the samples in W. The similarity between a new sample x and W is defined as Note that a smaller di v(x, W) value implies more diversity between x and W. In order to avoid selecting highly redundant samples, a novel diversity criterion is defined by combining the similarity between a query sample and the labeled set, and the similarity between any two candidate query samples at the same time. This calculation is given bywhere div(w , W) represents the diversity between the sample w and the candidate set W (apart from w ), and div(w , L) represents the diversity between the sample w and the labeled training set L.

3.3. Proposed D-AL-ELM Algorithm

The BvSB sampling method is a highly effective strategy for sample selection in active learning. Unfortunately, the BvSB may also select some uncertain samples which contain highly redundant information, which reduces the information available for classification. To address this problem, optimal samples were selected for classification. An ideal sample would not only furnish significant information for the classifier but also show diversity from the candidate unlabeled set and a minimal amount of redundancy within the labeled set. The specific steps for each iteration of the D-AL-ELM algorithm are as follows:where λ is a tradeoff parameter that can balance the informativeness and representativeness criteria, and L is the labeled training set. denotes the unlabeled sample that will be annotated and then included into the labeled training dataset for updating the ELM classifier. Step 1: the BvSB strategy is adopted to select the h most uncertain samples from the unlabeled samples pool U. Step 2: let h represent the most uncertain samples, denoted by W = {w 1, w 2,…, w } ⊆ U, and S be an arbitrary subset containing m(m ≤ h) samples selected from W. Two evaluations are then performed, including the diversity from the labeled set L and the candidate set S , and the similarity to the samples in S . Step 3: combining the discriminative and representative parts, the following formulation is obtained to select the m samples which are uncertain and diverse from each other: The implementation of the proposed method is summarized in Algorithm 1.
Algorithm 1

The double-criteria active learning with the ELM algorithm.

In order to quantitatively evaluate the quality of each learning algorithm, area under the learning curve (ALC) [13] was calculated as a performance metric, which is described aswhere N denotes the number of learning iterations and y denotes the classification accuracy at the i-th learning round, such that ALC ∈ [0,1]. It is noted that the larger the ALC value, the better the performance of the learning algorithm.

4. Experimental Results and Discussions

In this section, several experiments were performed on benchmark datasets and multiclass MI EEG datasets to evaluate the performance of the proposed D-AL-ELM method, in comparison with the other state-of-the-art approaches, including passive learning-based ELM, AL-ELM [13], and entropy-based ELM [14]. All methods were implemented using the MATLAB 2014b environment on a computer with a 2.5 GHz processor and 4.0 GB RAM.

4.1. Experiments on the Benchmark Datasets

4.1.1. Description of the Benchmark Datasets

A series of experiments were performed to evaluate the D-AL-ELM algorithm on 9 benchmark datasets from the KEEL dataset [29] and UCI dataset repositories [30]. Datasets included both binary and multiclass classification problems. As in [13], each raw dataset was divided into three parts: a small initial labeled set, a large unlabeled set, and a testing set. Testing instances comprised 50% of the total number of samples, while the percentage of initially labeled instances was assigned based on the size of the raw dataset and the number of categories. Detailed information regarding these datasets is presented in Table 1.
Table 1

Details of the datasets including the numbers of the corresponding features and samples.

DatasetNumber ofPercentage of initial labeled instances (%)Percentage of initial unlabeled instances (%)Percentage of test instances (%)
FeaturesInstancesClasses
Liver73452104050
Diabetes87682104050
Wdbc305692104050
Twonorm207400214950
Hayes-Roth41603104050
Iris41503104050
Wine131783104050
Segment1923107104050
Letter16200002614950

4.1.2. The Compared Algorithms, Parameter Settings, and the Performance Metric

In our experiments, we compare the proposed method with other state-of-the-art learning algorithms, including the following: PL-ELM: a passive learning algorithm that randomly selects some instances from the unlabeled set to train the initial classifier AL-ELM: a batch-mode active learning method based on ELM that uses the margin sampling strategy to select most uncertain examples for labeling [13] ELM-Entropy: querying discriminative samples through entropy measures [14] In this study, the ELM adopted a sigmoid function as the activation function on the hidden level. A grid search based on tenfold cross-validation was then used to find the optimal number of hidden nodes M in the initial labeled set. For the regularization parameter c, a leave-one-out (LOO) cross-validation strategy was adopted based on the minimum MSE to find the optimal parameter value [31]. The optimal parameters M and c were determined from M ∈ {10,20,…, 200} and c ∈ {e −5, e −4.9,…, e 5} on all the datasets except for the Letter dataset, where the parameter M was searched among {100,200,…, 1000}. Additionally, the tradeoff parameter λ ∈ {0.1, 0.2,…, 0.9} for equation (10) was chosen by grid search when M and c were fixed through the aforementioned methods. It should be noted that the ELM parameter selection process was implemented in the same manner for all four methods. Parameter details are shown in Table 2. It should be noted that the regularization parameter c was automatically identified using the LOO cross-validation and was not fixed during the learning process (thus, not shown in Table 2).
Table 2

Details of the optimal parameter settings for the different datasets using four methods.

DatasetD-AL-ELMAL-ELMELM-EntropyPL-ELM
M λ M M M
Liver1100.1110110110
Diabetes1100.3110110110
Wdbc2000.1200200200
Twonorm1200.1120120120
Hayes-Roth1000.3100100100
Iris1700.9170170170
Wine600.7606060
Segment2000.6200200200
Letter7000.4700700700
The batch mode was adopted to add new labeled instances. For the proposed D-AL-ELM method, h samples were first selected from the unlabeled set using equation (7), and then m samples were selected from the h samples using equation (11) and added to the labeled set for each iteration. In this experiment, h was empirically set to h=5m while m was 5% of the total instances in the original unlabeled set for 8 of the 9 datasets (except Letter). For the Letter dataset, m was 1% of the total instances in the original unlabeled set and h was set to h=2m. These parameters were chosen to decrease the labeling cost, considering the size of the raw dataset and the number of categories. To provide a fair comparison, all four methods queried m instances on each iteration. For each dataset, the procedure was stopped when the prediction accuracy stabilized or the number of selected samples was greater than 80% of the original unlabeled set. Additionally, to ensure the validity of experimental results, ten runs were performed for each learning method in each experiment, and average results were calculated.

4.1.3. Comparisons with Relevant State-of-the-Art Algorithms

Figure 1shows the trends of classification accuracy for the classifiers when trained by increasing numbers of data points across the various datasets. The results show that the proposed D-AL-ELM algorithm yielded the highest accuracy of all four methods on most of datasets (excepting the Wine and Iris datasets) at the last learning round. Specifically, the proposed method performed better than the remaining three methods over the majority of the active learning period for the Twonorm, Hayes-Roth, and Letter datasets. Moreover, the D-AL-ELM yielded the fastest learning rate over the first few iterations of the learning process for most datasets. This phenomenon indicates that the new method begins by effectively identifying the most informative and representative samples, unlike the other algorithms. Additionally, the ELM-Entropy approach generally yielded lower accuracy in multiclass classification, failing to surpass the PL-ELM on the Wine, Hayes-Roth, Iris, and Letter datasets. Another interesting observation was that the performance tended to degrade at a certain interval on the Segment dataset. It was considered that the Segment dataset may have a more irregular data structure, confounding the BvSB strategy and deteriorating the result. In cases such as this, a more adaptive stop criterion should be designed to stop the learning program at a more appropriate right time, before output degrades.
Figure 1

The learning curves of the four different learning algorithms on 9 benchmark datasets. (a) Liver. (b) Diabetes. (c) Wdbc. (d) Twonorm. (e) Wine. (f) Hayes-Roth. (g) Iris. (h) Segment. (i) Letter.

Table 3 presents the mean classification accuracies of the four methods across the 9 datasets during the learning process. The ALC values for the four methods are further compared in Table 4. The results shown in Tables 3 and 4 indicate that the D-AL-ELM method yielded the best performance among all datasets for the tested methods. As in [13], the ALC metric not only was related to the learning velocity but also had close relationship to the quality of the learning model. The proposed D-AL-ELM outperformed the other methods in terms of ALC, with the AL-ELM performing second best, with an accuracy close to that of the D-AL-ELM on the Wdbc and Segment datasets. For the Wdbc dataset, although the proposed method had a slightly higher ALC value than the AL-ELM, both algorithms yielded the same mean accuracy for the overall learning process.
Table 3

Mean accuracy results of the learning processes on 9 datasets (%).

DatasetD-AL-ELMAL-ELMELM-EntropyPL-ELM
Liver 67.59 67.4667.1466.14
Diabetes 76.31 76.1376.1274.60
Wdbc 96.18 96.18 96.1194.69
Twonorm 97.38 97.2597.2597.13
Wine 96.08 95.7295.6495.79
Hayes-Roth 59.43 56.2354.0356.56
Iris 96.65 96.4394.4495.58
Segment 89.26 89.1788.0687.63
Letter 82.89 81.6767.0975.76
Table 4

ALC comparisons of four methods on 9 datasets.

DatasetD-AL-ELMAL-ELMELM-EntropyPL-ELM
Liver 0.7624 0.76100.75720.7447
Diabetes 0.7640 0.76240.76220.7469
Wdbc 0.9624 0.96230.96160.9474
Twonorm 0.9742 0.97290.97290.9715
Wine 0.9622 0.95840.95730.9584
Hayes-Roth 0.5959 0.56220.53950.5663
Iris 0.9676 0.96550.94500.9563
Segment 0.8939 0.89290.88120.8766
Letter 0.8330 0.82040.67180.7606
Finally, Table 5 reports the average time for the learning stage of each algorithm across all datasets. As expected, the PL-ELM was the fastest method because it lacked any criteria for the evaluation of samples. The proposed D-AL-ELM required slightly more learning time than AL-ELM and ELM-Entropy, since it computed both informativeness and representativeness of each instance. Considering the improvement of classification performance, this extra time may be deemed an acceptable tradeoff.
Table 5

Average running time (s) for each learning algorithm.

DatasetD-AL-ELMAL-ELMELM-EntropyPL-ELM
Liver0.91410.75310.77190.7453
Diabetes1.20311.04381.00600.9719
Wdbc1.34841.25001.28281.2234
Twonorm7.98135.30475.58754.8844
Wine0.53910.46250.46250.4391
Hayes-Roth0.51090.45160.45940.4203
Iris0.52500.44690.48560.4313
Segment3.65783.390633123.3250
Letter121.8047115.5203123.0641111.4641

4.1.4. Analysis of Effect of Different Batch Size Values

In this experiment, the performance of the proposed active learning method was further evaluated using different batch sizes (i.e., h and m values). The new method was tested with different querying sizes by varying the values of h and m, respectively. The remaining experimental settings were the same as in earlier experiments and testing was conducted on two benchmark datasets: Iris and Wine. The M and λ parameters were set as M=100,  λ=0.5 to observe the performance with different batch sizes. Results are reported in Figures 2 and 3. In Figure 2, m was fixed at 5% of the total number of instances in the original unlabeled set and h was chosen from a candidate set {h=1.1m, 1.2m, 1.5m, 2m, 4m, 5m}. In Figure 3, h was fixed at the value of 2m and m was chosen from {1%, 2%, 4%, 6%, 8%}. It can be seen from Figure 2 that learning rates at the start of the curve increased with higher h values. Performance on Iris was less sensitive to the h value when enough instances were queried, and learning curves tended to be similar when query numbers and h values were large. In contrast, performance on the Wine dataset was more sensitive to h. This may be a result of the Wine dataset having a more complex distribution, which is difficult to capture. Although the D-AL-ELM performed differently on the two datasets, relatively larger h values were consistently able to obtain favorable performance. On the other hand, this increase in h value leads to a greater computational burden. Figure 3 shows the effects of different values of m on the Iris and Wine datasets. From this, it was observed that convergence can be more easily achieved with small m values. Alternatively, when m is large, more instances can be learned at each iteration and the number of total iterations greatly reduced, although this boost in performance does not provide substantially increased accuracy. In conclusion, optimizing the h and m values is not crucial for the D-AL-ELM, as most values yield similar results. It should be noted, however, that larger h and m are generally recommended.
Figure 2

The learning curves of the proposed algorithm with different h values on Iris and Wine. (a) Iris. (b) Wine.

Figure 3

The learning curves of the proposed algorithm with different m values on Iris and Wine. (a) Iris. (b) Wine.

4.2. Experiment on Multiclass MI EEG Data

4.2.1. Description of EEG Datasets

This section further evaluates the performance of the proposed D-AL-ELM method on multiclass MI EEG data from the BCI Competition IV Dataset 2a [32]. This dataset consists of the EEG signals from 9 subjects who performed 4 tasks, including left hand, right hand, foot, and tongue MI. EEG signals were recorded using 22 electrodes. Each subject underwent a training and testing session, each consisting of 288 trials (a total of 576 trials across the two sessions).

4.2.2. Experimental Setup and Parameter Settings

Data preprocessing was first performed on the raw EEG data. For each trial, features were extracted from the time segment lasting from 0.5 s to 2.5 s after the cue instructing the subject to perform MI. Each trial was first band-pass filtered from 8–30 Hz using a fifth-order Butterworth filter. Next, the dimension of the EEG signal was reduced to a 24-dimension feature set using the one-versus-rest common spatial pattern (OVR-CSP) algorithm [33], which is an effective and popular feature extraction method for EEG multiclassification that computes the features that discriminate each class from the remaining classes. Finally, the features extracted by OVR-CSP were discriminated using the different classification methods. Optimal selection of the M, λ, and c parameters was performed in the same manner described in Section 4.1.2. The number of hidden nodes M was searched within {10,20, ..., 150}. For each subject, the first 400 trials were considered as the training set, while the remaining 176 trials were used as the independent testing set [11]. The values for m and h were set at m=10 and h=5m. Finally, experiments included ten runs for each learning method from which average results were calculated.

4.2.3. Comparisons with Related Algorithms

Figure 4 illustrates the trend lines of classification accuracy when methods were applied to different testing datasets, while Table 6 lists the mean classification accuracies of the four methods during the learning process. Table 7 then provides the ALC results, while Table 8 shows the average running time (s) for the learning stage.
Figure 4

Learning curves of the four different learning algorithms on BCI Competition IV Dataset 2a. (a) S1. (b) S2. (c) S3. (d) S4. (e) S5. (f) S6. (g) S7. (h) S8. (i) S9.

Table 6

Mean accuracy (%) of the learning process on BCI Competition IV Dataset 2a.

DatasetsD-AL-ELMAL-ELMELM-EntropyPL-ELM
S1 83.78 83.6483.3283.76
S253.9152.5052.33 54.55
S385.66 86.06 85.5784.39
S4 65.32 63.4862.9464.95
S5 49.06 48.9047.7948.00
S6 54.31 53.9652.9752.36
S783.15 83.34 82.4581.96
S883.46 83.56 83.3481.71
S9 83.57 82.8082.3982.91
Mean 71.36 70.9270.3470.51
Table 7

ALC values of the four methods on BCI Competition IV Dataset 2a.

DatasetsD-AL-ELMAL-ELMELM-EntropyPL-ELM
S1 81.22 0.81090.807681.21
S20.52380.51000.5085 0.5296
S30.8302 0.8340 0.82910.8177
S4 0.6340 0.61590.61050.6308
S5 0.4764 0.47490.46380.4662
S6 0.5276 0.52410.51440.5088
S70.8066 0.8085 0.79960.7952
S80.8090 0.8100 0.80790.7923
S9 0.8103 0.80320.79920.8042
Table 8

Average running time (s) of each learning algorithm.

DatasetsD-AL-ELMAL-ELMELM-EntropyPL-ELM
S11.72661.46251.45941.4172
S22.51252.28132.38592.2469
S31.72191.48591.45781.4234
S42.07191.78911.83281.8125
S52.07811.82811.81411.7719
S61.76091.45941.40941.3609
S72.41882.19692.17342.1641
S81.55621.33751.36091.3078
S91.29060.94840.95630.8906
Mean1.90421.64321.65001.5995
The results show that the performance of D-AL-ELM method is comparable to that of the AL-ELM and better than that of the ELM-Entropy and PL-ELM algorithms for all subjects (except for subject 2 in PL-ELM). Specifically, the proposed method surpassed the AL-ELM approach in 6 of the 9 subjects (1, 2, 4, 5, 6, 9) in terms of the ALC metric. For all 9 subjects, the D-AL-ELM method yielded a mean accuracy of 71.36%, higher than that of AL-ELM (70.92%), ELM-Entropy (70.34%), and PL-ELM (70.51%). These results demonstrate the effectiveness of the D-AL-ELM in selecting both informative and representative instances from unlabeled EEG samples. Additionally, they reveal that the proposed method can calibrate an effective classifier for MI EEG signals without the need for a large number of labeled training samples. For comparative purposes, Table 8 also provides the average running time of each learning algorithm. Although the D-AL-ELM exhibited slightly longer training time than the other three methods, this may be considered a worthwhile tradeoff for the improved classification performance of the D-AL-ELM.

4.3. Discussion

In these experiments, the proposed D-AL-ELM method exhibited excellent performance in both classification accuracy and computational efficiency, as demonstrated on several benchmark datasets and an experimental MI EEG dataset. When compared to a passive learning-based ELM, D-AL-ELM achieved improved performance by effectively extracting the most valuable unlabeled samples. The D-AL-ELM also outperformed the AL-ELM and ELM-Entropy algorithms, which both employed a single-query strategy. Improvement was seen on all nine datasets in Section 4.1, evidencing the ability of the D-AL-ELM to boost overall learning performance by combining the uncertainty and diversity strategies when updating the classifier with the selected samples. In terms of computational efficiency, the slight increase in training time for the D-AL-ELM, as compared to the PL-ELM, AL-ELM, and ELM-Entropy, was negligible in practice, especially when considering the improved classification accuracy. The experimental results then demonstrate that the proposed algorithm can effectively and comprehensively measure the representativeness of samples. Simultaneously, the proposed approach also measures how informative individual examples are, contributing to the improved classifier performance. Combining these factors, suitable instances can be selected for classifier construction. Finally, the effectiveness of the D-AL-ELM was shown in its application to an experimental multiclass MI task from the BCI Competition IV Dataset 2a. Due to the low signal-to-noise ratio of EEG data, the applied algorithms struggled to generate adequate results. Consequently, hand-designed features were first extracted from the raw EEG data using the OVR-CSP approach, and the different AL algorithms were then used to further extract the unlabeled samples and calibrate a robust classifier. For subjects S1, S3, S7, S8, and S9, the D-AL-ELM yielded an acceptably high mean classification accuracy of over 80% for the whole learning process. Unfortunately, all tested methods performed poorly on subject S5. The proposed algorithm was only able to achieve 49.06% accuracy which, though insufficient, still ranked the highest among the applied methods.

5. Conclusion

In this paper, a novel active learning method with ELM, the D-AL-ELM, was developed for multiclassification. This new algorithm combines the uncertainty and diversity strategies and effectively reduces the expense and time cost of obtaining labeled data manually. For each sample, the proposed algorithm employs a BvSB strategy to measure informativeness and the cosine angle distance to measure diversity. The modified diversity measure not only estimates the diversity between the limited labeled training data and previously selected unlabeled samples, but also calculates the similarity among previously selected samples. Experimental results from several benchmark datasets and the multiclass MI EEG data from BCI Competition IV Dataset 2a were then used to verify the efficacy of the proposed D-AL-ELM algorithm. These results indicate that the performance of the proposed algorithm is consistently better than, or at least comparable to, that of other popular active learning techniques. Future work will then aim to develop online learning for the D-AL-ELM [23, 34]. In addition, an adaptive stopping criterion may be applied to promote the efficiency of the D-AL-ELM and improve its abilities for the classification and evaluation of MI EEG signals.
  17 in total

1.  A Maximum Entropy Framework for Semisupervised and Active Learning With Unknown and Label-Scarce Classes.

Authors:  Zhicong Qiu; David J Miller; George Kesidis
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2016-01-26       Impact factor: 10.451

Review 2.  EEG-Based Strategies to Detect Motor Imagery for Control and Rehabilitation.

Authors:  Kai Keng Ang; Cuntai Guan
Journal:  IEEE Trans Neural Syst Rehabil Eng       Date:  2016-12-30       Impact factor: 3.802

3.  Scalable active learning for multiclass image classification.

Authors:  Ajay J Joshi; Fatih Porikli; Nikolaos P Papanikolopoulos
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2012-11       Impact factor: 6.226

4.  Extreme learning machine and adaptive sparse representation for image classification.

Authors:  Jiuwen Cao; Kai Zhang; Minxia Luo; Chun Yin; Xiaoping Lai
Journal:  Neural Netw       Date:  2016-06-23

5.  Active Learning by Querying Informative and Representative Examples.

Authors:  Sheng-Jun Huang; Rong Jin; Zhi-Hua Zhou
Journal:  IEEE Trans Pattern Anal Mach Intell       Date:  2014-10       Impact factor: 6.226

6.  Exploring Representativeness and Informativeness for Active Learning.

Authors:  Bo Du; Zengmao Wang; Lefei Zhang; Liangpei Zhang; Wei Liu; Jialie Shen; Dacheng Tao
Journal:  IEEE Trans Cybern       Date:  2015-11-17       Impact factor: 11.448

7.  Review of the BCI Competition IV.

Authors:  Michael Tangermann; Klaus-Robert Müller; Ad Aertsen; Niels Birbaumer; Christoph Braun; Clemens Brunner; Robert Leeb; Carsten Mehring; Kai J Miller; Gernot R Müller-Putz; Guido Nolte; Gert Pfurtscheller; Hubert Preissl; Gerwin Schalk; Alois Schlögl; Carmen Vidaurre; Stephan Waldert; Benjamin Blankertz
Journal:  Front Neurosci       Date:  2012-07-13       Impact factor: 4.677

8.  Scale-Dependent Signal Identification in Low-Dimensional Subspace: Motor Imagery Task Classification.

Authors:  Qingshan She; Haitao Gan; Yuliang Ma; Zhizeng Luo; Tom Potter; Yingchun Zhang
Journal:  Neural Plast       Date:  2016-11-03       Impact factor: 3.599

9.  Multiclass Informative Instance Transfer Learning Framework for Motor Imagery-Based Brain-Computer Interface.

Authors:  Ibrahim Hossain; Abbas Khosravi; Imali Hettiarachchi; Saeid Nahavandi
Journal:  Comput Intell Neurosci       Date:  2018-02-22

10.  Enhancing Performance of a Hybrid EEG-fNIRS System Using Channel Selection and Early Temporal Features.

Authors:  Rihui Li; Thomas Potter; Weitian Huang; Yingchun Zhang
Journal:  Front Hum Neurosci       Date:  2017-09-15       Impact factor: 3.169

View more
  1 in total

1.  EEG-Based Driving Fatigue Detection Using a Two-Level Learning Hierarchy Radial Basis Function.

Authors:  Ziwu Ren; Rihui Li; Bin Chen; Hongmiao Zhang; Yuliang Ma; Chushan Wang; Ying Lin; Yingchun Zhang
Journal:  Front Neurorobot       Date:  2021-02-11       Impact factor: 2.650

  1 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.