Literature DB >> 30463202

Gas-Sensor Drift Counteraction with Adaptive Active Learning for an Electronic Nose.

Tao Liu1, Dongqi Li2, Jianjun Chen3, Yanbing Chen4, Tao Yang5, Jianhua Cao6.   

Abstract

Gas sensors are the key components of an electronic nose (E-nose) in violated odour analysis. Gas-sensor drift is a kind of physical change on a sensor surface once an E-nose works. The perturbation of gas-sensor responses caused by drift would deteriorate the performance of the E-nose system over time. In this study, we intend to explore a suitable approach to deal with the drift effect in an online situation. Considering that the conventional drift calibration is difficult to implement online, we use active learning (AL) to provide reliable labels for online instances. Common AL learning methods tend to select and label instances with low confidence or massive information. Although this action clarifies the ambiguity near the classification boundary, it is inadequate under the influence of gas-sensor drift. We still need the samples away from the classification plane to represent drift variations comprehensively in the entire data space. Thus, a novel drift counteraction method named AL on adaptive confidence rule (AL-ACR) is proposed to deal with online drift data dynamically. By contrast with conventional AL methods selecting instances near the classification boundary of a certain category, AL-ACR collects instances distributed evenly in different categories. This action implements on an adjustable rule according to the outputs of classifiers. Compared with other reference methods, we adopt two drift databases of E-noses to evaluate the performance of the proposed method. The experimental results indicate that the AL-ACR reaches higher accuracy than references on two E-nose databases, respectively. Furthermore, the impact of the labelling number is discussed to show the trend of performance for the AL-type methods. Additionally, we define the labelling efficiency index (LEI) to assess the contribution of certain labelling numerically. According to the results of LEI, we believe AL-ACR can achieve the best effect with the lowest cost among the AL-type methods in this work.

Entities:  

Keywords:  active learning; drift counteraction; electronic nose; online

Year:  2018        PMID: 30463202      PMCID: PMC6263697          DOI: 10.3390/s18114028

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.576


1. Introduction

An electronic nose (E-nose) is a potential approach to perform odour analysis. A typical E-nose system ordinarily consists of two parts: a gas-sensor array and a pattern-recognition unit. The former part generates odour fingerprints from gas-sensor responses with high cross sensitivity while the latter performs data analysis using computational algorithms. With this structure, E-noses can identify complicated gas components with low-cost gas-sensor arrays. Owing to the advantages in cost and operation, E-noses have been adopted in environmental monitoring [1], the food industry [2], agriculture [3] and medicine [4]. The modern E-nose system has appeared as an artificial intelligence machine since the 1990s [5]. There are several scientific fields, such as artificial intelligence, system integration and sensor technology supporting the development of E-noses. Gas-sensor drift in E-nose systems is a kind of concept drift, which frequently occurs by surface aging, environmental disturbance and sensor poisoning. This phenomenon deteriorates the compatibility between gas-sensor responses and algorithm models over time, and finally causes the performance degradation of E-noses. In other words, the algorithm models would be meaningless without any drift calibrations over time. Thus, anti-drift methods are necessary for E-nose systems in their life-long working process. To reduce the drift effect of E-noses, possible solutions have been focused on sensor improvement [6] and algorithm modification [7]. According to current studies, algorithm-modification approaches have received increasing attention due to the achievements of artificial intelligence and machine learning. Signal pre-processing and classifier updating are two mainstreams in algorithm modification. The former manner extracts drift-like component from original gas-sensor responses and then restructures responses without drift. Huang et al. adopted principal component analysis (PCA) and common PCA (CPCA) at different gas concentrations to compensate drift signals [8]. Ziyatdinov et al. used CPCA to discover a drift direction for all gasses in feature space [9]. Besides PCA, orthogonal signal correction (OSC) [10] and independent component analysis (ICA) [11] are alternatives for drift counteraction in the time domain. Furthermore, the wavelet is another way to remove drift component in time-frequency domain [12]. In terms of classifier updating, the classifier ensemble is a popular strategy for long-term drift migration [13,14,15]. This strategy provides a weighted result based on several sub-classifiers generated from different samples. Moreover, some scholars used adaptive classifiers including self-organize maps (SOM) [16,17], adaptive resonance theory (ART) [18,19] and immune algorithm [20], to handle drift distribution adaptation dynamically. Recently, the algorithm modification manner has been extended to solve the drift problem in semi-supervised learning situations. Liu et al. put a transfer learning paradigm into a semi-supervised scene for drift elimination [21]. All the above methods have a common premise that a calibration set with complete categories and well-labelled samples should be prepared for drift correction. However, this demand would be barely satisfied in online scenarios due to inadequate time and labour to obtain such a calibration set, for example, in online toxic-gas monitoring. In this application, toxic gases rare appear while non-hazardous gases are majority. Therefore, the collected calibration set is imbalanced, which would decline the performance of the E-nose in toxic-gas recognition. Additionally, we assume that the toxic-gas recognition would be a failure according to the unreliable labels of calibration samples from the classifiers. In this study, we use active learning (AL) to solve the calibration-set problem. AL methods can select a certain number of instances from recent incoming samples, then label the selected instances and update classifiers reliably. We believe this working process can adapt well to online scenarios due to the fact that AL methods have the ability to select relatively optimal instances from undetermined calibration samples for classifier upgrading. Moreover, conventional AL instance-selection strategies select instances with low confidence, which cannot represent the current distribution of drifting samples entirely. Consequently, we have proposed an innovative AL method on the adaptive confidence rule (AL-ACR) to balance the categories of selected instances. It is helpful to enhance the uniformity of drift information in a calibration set. We use two E-nose drift databases as benchmarks to evaluate the performance of the proposed method in drift counteraction. One dataset is public and the other was collected by the authors. According to the experimental results, we infer that the proposed AL-ACR is suitable for online drift calibration. The AL-ACR shows adequate adaptability and feasibility in online scenarios and reaches better accuracy comprehensively than other references. Considering the time-consuming and laborious requirements of manual labelling, it is crucial to obtain a higher accuracy with a lower labelling number. We therefore define the labelling efficiency index (LEI) to assess the efficiency of AL methods. The results of LEI show that AL-ACR has obvious advantages in most cases. In this paper, we explain the development of AL in Section 2. Section 3 describes the details of the proposed AL method. In Section 4, we show and discuss the comparative results of the proposed and reference methods upon two databases. Finally, we summarize some conclusions from the experimental results in Section 5.

2. Related Work

The AL paradigm generally consists of two modules, a selecting engine and a learning engine [22]. The selecting engine is designed to achieve high-quality unlabelled samples according to the classifiers constructed by the learning engine. It labels the selected samples manually and adds the labelled instances to the training set. The learning engine is responsible for renewing classifiers by the both training set and the manual labelled instances, which makes the performance of the classifier improve continuously. The cycle of “selecting-learning” would be stopped if certain termination condition were reached. It is obvious that how to select high-quality instances from unlabelled samples on feature information is a key part of the AL method. It plays an important role in promoting the classification performance of AL. Therefore, studies on instance-selection strategies have gradually increased. Membership query synthesis (MQS), stream-based and pool-based methods are typical methods proposed to select unlabelled samples. MQS is the earliest idea of AL using a query to learn [23]. In this method, the hypothetical learning system can ask questions from experts, that is, the MQS determines labels of certain instances by query. As for the instance selection of MQS, all unlabelled samples are handed to the expert for labelling, regardless of labour cost and data distribution. To avoid unnecessary labelling, scholars have proposed stream-based AL which selects instances according to either information entropy or distribution similarity and labels them in an instance-by-instance manner [24]. Although the stream-based AL has solved the drawbacks of the MQS to a certain extent, it still demands a fixed threshold to measure the information contained in each input sample. Thus, unsteady labelling density and rigid parameter adaptation would limit the performance of stream-based AL. Consequently, Lewis et al. proposed a pool-based AL method to form a “pool” for expert labelling [25]. This pool-based AL selectively labels instances that exist in a dataset to enhance the performance of classifiers. In fact, this method becomes equivalent to the stream-based AL when just one unlabelled sample exists in the pool. In other words, the stream-based AL is a special case of the pool-based one. For the instance-selection methods of pool-based AL, how to select the most representative instances from the candidates in order to label is crucial. Uncertainty sampling (US) and query-by-committee (QBC) strategies are main approaches for instance selection. The former identifies samples on distributions while the latter selects instances according to the outputs of classifiers. Recently, in the studies of E-noses, Jiang X et al. proposed an enhanced QBC (EQBC) -radial basis function neural network (RBFNN) [26] for the recognition improvement on the QBC strategy. However, there are hardly any pubic reports for drift counteraction of E-noses with active learning. As far as we know, this study is the first research on drift compensation of E-noses with AL methods.

3. Methods

3.1. Baseline Methods

There are three AL methods including US, QBC and EQBC presented as baseline for comparison. Considering the various variants of them, we use US proposed in Ref. [27], QBC and EQBC in Ref. [26]. The US strategy uses the classifiers to identify the largest indeterminacy of unlabelled samples for subsequent labelling. The focus of US is to measure the uncertainty of the sample. It assumes the most uncertain sample contributes to the improvement of classification at most. We can obtain the optimal sample for recognition improvement by calculating posterior probability. We compute the margin of instance p as follow:where and are, respectively, the class with the maximum posteriori probability and the second most posteriori probability. The margin-based metric is prone to select instances with minimum margin between posteriori probabilities of the two most likely class labels. The QBC method is the most popular AL method based on version space reduction. It aims to select samples in a simplified space for man-made labelling, which has been proposed by Seung et al. [28] and Freund et al. [29]. The QBC sets up a committee group to vote a sample with the highest disagreement for expert labelling. We choose Kullback–Leibler (KL) [30] as the metric for QBC to measure the valuable instances for classification promotion. The calculation formula of KL can be given by Equations (1) and (2) as follows: where represents the probability that sample p is labelled as Class c by member k and K denotes the total number of committee members. The EQBC method is based on a weighted combination of Kullback–Leibler divergence (KL−d) and vote entropy (VE) as shown in Equation (4). where ω1 and ω2 are two adjustable weighted parameters.

3.2. Adaptive Instance Selection

Baseline AL methods regularly select useful instances with a single criterion with either high information entropy or low similarity, which is suitable to static data due to the constant data distribution. The classification plane is more distinguished by the increasing number of selected instances, regardless of the instance expiration. Unfortunately, in online mode, instance expiration is bound to exist under drift influence. This limitation may lead AL methods based on single criterion to barely reflect current drift trend by category in time. Finally, in a period, the performance of one class may rise, whereas others would incline. Thus, we proposed an innovative pool-based AL method, AL-ACR, with different instance-selection criteria adaptively. There are two criteria used here to discover the instances evenly in distribution by category. We adopt information entropy to index the necessity of samples for labelling. A high information-entropy value shows low confidence, vice versa. As Figure 1 described, the AL-ACR method implements instance selection and classifier promotion iteratively. F is the index of min/max confidence and its default value equals 0. F = “0” denotes that the minimum confidence should be used while F = “1” infers that the maximum confidence should be selected. For classifiers, the chairman classifier 0 is trained by the whole training set, and the member classifiers 1 and 2 are generated from parts of the training set randomly. In an instance-selection process, we primarily compare the chairman outputs of the current instance before and after classifier updating. On the one hand, if they are different, it means the classification plane of the chairman needs to be distinguished, then the sample with maximum information entropy to specify the classification boundary should be selected in the next round. On the other hand, if they are the same, we assume the category of current instance is well classified, then instances belonging to other kinds should be designated with minimum information entropy according to label diversity and distribution equilibrium. For classification promotion, when entering a new instance, the corresponding label is assigned by expert opinion. After that, both the new instance and corresponding label are added to the training sets belonging to the members of the committee. Next, the members are respectively retrained by their training sets.
Figure 1

Flow chart of active learning on adaptive confidence rule (AL-ACR) algorithm.

4. Experiments and Results

4.1. Datasets

4.1.1. Public Dataset

The public dataset (http://archive.ics.uci.edu/ml/machine-learning-databases/00224/) is from the machine learning repository of University of California Irvine [31]. We denote this dataset as Dataset A for short in following discussion. There are totally 13,910 samples collected by an E-nose over three years (36 months). The gas-sensor array consists of sensor devices (4 of each) as TGS2600, TGS2602, TGS2610 and TGS2620; in total 16 gas sensors are adopted in the E-nose system. This dataset includes six-gas samples dosed at different concentrations. The entire response of a gas sensor in an experiment can be divided into three phases: injection, measurement and sweeping. Consequently, 8 features (2 steady-state features and 6 transient features) have been abstracted from these three phases for each sensor. On account of the sensor array made up of 16 gas sensors, one experiment can collect 128 (8 × 16) features. Thus, the total size of Dataset A is 13,910 × 128. All 10 batches have been arranged by time order. To reduce the imbalance of batch sizes, we integrated Batch 4 and 5 as Batch 4&5, Batch 8 and 9 as Batch 8&9. Thus, we actually prepared 8 bunches of drifting samples for evaluation. Additionally, the working temperature of the gas sensors was maintained at 400 °C and the sampling rate was set to 100 Hz. Figure 2a–h shows the distributions of the 8-bunch samples in principal component analysis (PCA) plots. We can infer that the drift effect acts apparently on the samples in all batches. The distribution of data varies continuously with time, especially between different batches. We believe that classifiers worked with no drift counteraction would be invalid rapidly.
Figure 2

(a) Data distribution of Batch 1; (b) Data distribution of Batch 2; (c) Data distribution of Batch 3; (d) Data distribution of Batch 4&5; (e) Data distribution of Batch 6; (f) Data distribution of Batch 7; (g) Data distribution of Batch 8&9; (h) Data distribution of Batch 10.

4.1.2. Collected Dataset

Besides Data A, the other dataset was collected from an E-nose system we designed. This E-nose system consists of three parts: gas sensor array, control unit and upper computer. The gas-sensor array includes 32 gas sensors with solid electrolyte, electrochemical and metal oxide types. The control unit transfers the fingerprints from the gas-sensor array to the upper computer for further processing. We gathered 63, 189 and 189 samples in the first, third and fourth months, respectively. Consequently, we divided the 4-month dataset into three batches according to chronological order and each batch had all kinds of samples equally in quantity. For each sample, we used the steady-state value of certain gas-sensor responses as a feature. Thus, one gas sensor corresponds one feature and one sample can be denoted as a 32-dimensional vector according to 32 gas sensors. Here, we define this dataset as Dataset B for following discussions. Figure 3a–c illustrate the distribution of Dataset B with PC1 and PC2 by batch.
Figure 3

(a) Data distribution of Batch 1 of Dataset B; (b) Data distribution of Batch 2 of Dataset B; (c) Data distribution of Batch 3 of Dataset B.

4.2. Experimental Setup

Our evaluation contains four parts. We primarily evaluated the accuracy of proposed AL-ACR compared with those of other start-of-the-art methods (AL-US, AL-QBC, AL-EQBC, RBF-SVM, GFK-SVM, Comgfk-SVM, Comgfk-ML and RBF-ML [21]) in long-term online drift counteraction. Secondly, we defined 4 different long-term scenarios to assess the performance between AL methods with labelled-instance number, category balance and labelling efficiency. The defined Scenarios 1–4 are as follows: Scenario 1 (Dataset A): Batch 1 for initial training, Batch 2–10 for online testing. Scenario 2 (Dataset A): Batch 2 for initial training, Batch 3–10 for online testing. Scenario 3 (Dataset A): Batch 3 for initial training, Batch 4–10 for online testing. Scenario 4 (Dataset B): Batch 1 for initial training, Batch 2 and 3 for online testing. For each scenario, data pre-processing has been performed to normalize sample features in [0, 1] by:where x represents a 128-dimensional sample, is the output after pre-processing, and is the maximum and minimum feature value in x respectively. In terms of classifiers, we adopt three types: k-nearest neighbour (k-NN), support vector machine (SVM) and radial basis function neural network (RBFNN). The parameter k of k-NN is set to 3. For SVM, we adjust kernel function, kernel parameter σ (kernel function only) and penalty factor C to optimize the performance of SVM. During optimization of SVM, either radial basis or linear function could be selected. We set the scope of kernel parameter σ and the finest value of C to the range 10−3 to 103 with variable step. Considering the computational complexity and overall performance, we finally choose linear function as our option and set C = 0.2 (dataset A) or C = 2 (Dataset B). For RBFNN, we set the number of hidden layers to 50, the training error to 10−5 and the variance of radial basis kernel to 1. For AL methods, we chose roughly 30% of the data as pool samples per batch while the remains are used for testing only. We consider that it is impossible to label instances at any time in reality. Hence, the time of labelling should be concentrated in a period per batch. In the following section, on account of the fairness, the accuracies of the AL methods are obtained on whole batches including both pool and testing data in Table 1. Others are concluded on the testing data due to the comparison just between AL methods.
Table 1

Accuracy of methods in Scenario 1 (%).

RBF-SVMGFK-SVMComgfk-SVMComgfk-MLRBF-MLAL-USAL-QBCAL-EQBCAL-ACR
Batch 274.3672.7574.4780.2542.2553.5565.3465.23 98.61
Batch 361.0370.0870.1574.9973.6983.2665.8365.80 87.39
Batch 4&532.9668.6468.2072.5370.7057.9369.7669.76 72.64
Batch 628.2673.8273.9977.8277.5151.6244.1844.18 93.12
Batch 728.8154.5354.59 71.68 54.4360.0842.3842.3869.52
Batch 8&928.8064.16 64.71 50.4727.3966.2440.9640.9655.76
Batch 1034.4741.7853.7953.7934.9259.8837.9537.93 55.72
Average36.0955.7257.4960.1947.6154.0745.8045.78 66.60

4.3. Results and Discussion

4.3.1. Performance Evaluation of Paradigms

Table 1 has shown the accuracies of different paradigms in the Scenario 1. The supervised paradigm methods, RBF-SVM, GFK-SVM, Comgfk-SVM, are SVM with radial basis, geodesic flow and associated combined kernel, respectively. Another paradigm is semi-supervised learning including Comgfk-ML and RBF-ML. In terms of the AL paradigm, AL-US, AL-QBC, AL-EQBC and the proposed AL-ACR methods are used for comparison. We abstract the results of RBF-SVM, GFK-SVM, Comgfk-SVM, Comgfk-ML and RBF-ML directly from Ref. [21]. In this subsection, we calculate the results of AL methods in scenario 1 and use SVM only for performance comparison owing to the same settings adopted in Ref. [21]. We set the number of selected instance N to 6 for AL methods, a comparatively small number leading to the indistinctness of drift description. Reasonably, the performance of AL methods would be restricted to a poor level in the range of N. Among all the methods with different paradigms, AL-ACR has clearly reached the best recognition performance. The above phenomena indicate that the proposed method is suitable for online drift suppression in most cases. For the paradigm, we believe the supervised learning manner cannot compensate drift well for a long time; the semi-supervised learning methods are weak in robustness due to the unreliability of online labels.

4.3.2. Effect of Instance Number

The number of labelled instances N for each batch is an important parameter of AL methods. We investigate the recognition-rate shift of the AL-ACR models by increasing N value from small to large. In order to avoid class imbalance by the number of labelled instances, we set the increasing step of N to the category size of the dataset. Thus, for Scenarios 1–3, the step of N is 6 while that of Scenario 4 is 7. As Figure 4a–d, Figure 5a–d and Figure 6a–d show, the green line with cross, indigo line with diamond, blue line with star and red line with star represent the accuracy of AL-US, AL-QBC, AL-EQBC and AL-ACR, respectively. The recognition rates of three AL methods are presented. We can infer that larger N causes higher recognition rate in all cases. We believe the drift information has been described more abundantly and accurately since N increased. Compared with the other three methods, AL-ACR has achieved the best accuracy in most cases. This confirms the significant effectiveness of AL-ACR for online drift suppression of E-noses once again and the performance of reference methods is lagging behind AL-ACR. In other words, AL-ACR is an optimal choice almost under different parameters N from the results of Figure 4a–d, Figure 5a–d and Figure 6a–d.
Figure 4

(a) Accuracy of k-nearest neighbour (k-NN) (Scenario 1); (b) Accuracy of k-NN (Scenario 2); (c) Accuracy of k-NN (Scenario 3); (d) Accuracy of k-NN (Scenario 4).

Figure 5

(a) Accuracy of support vector machine (SVM) (Scenario 1); (b) Accuracy of SVM (Scenario 2); (c) Accuracy of SVM (Scenario 3); (d) Accuracy of SVM (Scenario 4).

Figure 6

(a) Accuracy of radial basis function neural network (RBFNN) (Scenario 1); (b) Accuracy of RBFNN (Scenario 2); (c) Accuracy of RBFNN (Scenario 3); (d) Accuracy of RBFNN (Scenario 4).

For both datasets A and B, the performance of algorithms has entered the bottleneck or even a certain decline since N became large. This is reasonable due to drift-information redundancy caused by the mismatch between N and the drift-information time scale. Such redundant information is meaningless to the AL models, and even lead to negative effects. Additionally, we use three classifiers, k-NN, SVM and RBFNN, to test whether the performance of the proposed methods has selectivity for classifiers. We note that all the classifiers share almost the same trend in terms of accuracy. On the one hand, SVM often shows stronger recognition ability than k-NN and RBFNN. This may be caused by the fact that SVM searches for optimal solutions globally whereas k-NN is based on local discriminate and RBFNN may suffer over-fitting. On the other hand, the performance difference of all four methods seems smaller on SVM than both k-NN and RBFNN. We infer that the excellent recognition ability of SVM may have offset the drawbacks of AL-US, AL-QBC and AL-EQBC.

4.3.3. Distribution of Labeled Instances

As previously described, the AL-ACR method tries to select instances evenly by category. This action would help the classifiers to capture entire drift details near and away from the classification boundaries during online drift compensation. As Figure 7 and Table 2 show, we present the true labels of selected instances with different colours in all scenarios. Black, ochre, blue and red dotted box denotes the bar of AL-US, AL-QBC, AL-EQBC and AL-ACR, respectively. The height of the bar denotes the instance quantity of a certain category. This allows us to see 4 bars in each scenario. Furthermore, we set total number of labelled instances to 210, 216, 210 and 70 for Scenarios 1–4 respectively. We can infer from the results that AL-ACR has the most balanced instance category to achieve a complete drift trend in online mode. The other three methods are slightly worse than AL-ACR in label balance. According to the ratios of instance classes in Table 2, we can find more balanced category distribution appeared in AL-ACR, which can properly explain why AL-US, AL-QBC and AL-EQBC perform weaker in terms of accuracy.
Figure 7

Label distribution in Scenarios 1–4.

Table 2

Category distribution in Scenarios 1–4 (selected number/ratio).

Class 1Class 2Class 3Class 4Class 5Class 6Class 7
SCENARIO 1 US 12/5.71%35/16.67%99/47.14%5/2.38%34/16.19%25/11.90%0/0.00%
QBC 61/28.24%56/25.93%18/8.33%48/22.22%24/11.11%3/1.39%0/0.00%
EQBC 60/28.57%57/27.14%16/7.62%48/22.86%21/10.00%8/3.81%0/0.00%
ACR 28/13.33%33/15.71%23/10.95%50/23.81%45/21.43%31/14.76%0/0.00%
SCENARIO 2 US 82/37.96%15/6.95%7/3.24%94/43.52%10/4.63%8/3.70%0/0.00%
QBC 28/12.96%33/15.28%98/45.37%13/6.02%36/16.67%8/3.70%0/0.00%
EQBC 16/7.41%28/12.96%111/51.39%20/9.26%36/16.67%5/2.31%0/0.00%
ACR 28/12.96%31/14.35%35/16.20%42/19.45%51/23.61%29/13.43%0/0.00%
SCENARIO 3 US 43/20.48%10/4.76%15/7.14%83/39.52%35/16.67%24/11.43%0/0.00%
QBC 1/0.46%61/28.24%53/24.54%0/0.00%85/39.35%10/4.63%0/0.00%
EQBC 1/0.48%53/25.24%60/28.57%5/2.38%80/38.10%11/5.23%0/0.00%
ACR 35/16.67%47/22.38%26/12.38%48/22.86%31/14.76%23/10.95%0/0.00%
SCENARIO 4 US 0/0.00%17/24.29%18/25.71%1/1.42%17/24.29%17/24.29%0/0.00%
QBC 18/25.71%14/20.00%4/5.71%4/5.71%9/12.86%6/8.57%15/21.44
EQBC 18/25.71%16/22.86%4/5.71%7/10.00%8/11.43%4/5.71%13/18.58%
ACR 9/12.86%12/17.14%14/20.00%8/11.43%11/15.71%13/18.57%3/4.29%

US: uncertainty sampling; QBC: query-by-committee; EQBC: enhanced QBC; ACR: adaptive confidence rule.

4.3.4. Values of Labelling Efficiency Index (LEI)

In general, we hope the AL methods can not only achieve high accuracy, but also reduce the number of labels to save cost in terms of time and laboriousness. In other words, raising the accuracy increment per labelling is crucial in AL. Therefore, we define LEI to demonstrate the cost of a certain AL method in recognition improvement. The LEI is a hybrid index as follows:where Acc is the accuracy under the number of selected instances N, ΔAcc is the increment of accuracy since the last labelling and α is a adjust parameter belongs to [0, 1]. We define LEI as a sum of two terms that represent current performance and associated increment, respectively. A qualified AL method should have excellent classification ability and economic cost of performance enhancement simultaneously. We set α to 0.2 for LEI computation. Table 3, Table 4, Table 5, Table 6, Table 7 and Table 8 illustrate the LEI values in all scenarios on the three classifiers with different N. In Scenario 1, no matter either SVM, k-NN or RBFNN, AL-ACR reaches the highest LEI score from beginning to end. In Scenario 2, AL-ACR is still the winner in all cases on both k-NN and RBFNN while 7 out of 10 cases on SVM are best with AL-ACR. As for Scenarios 3 and 4, although the performance of AL-ACR sometimes falls down compared with other references, it still keeps the highest LEI values in most cases. In a word, the AL-ACR has the higher labelling efficiency compared with AL-US, AL-QBC and AL-EQBC in different scenarios almost. We assume that AL-ACR is a very economical and practical approach to use for drift counteraction of E-noses in online mode.
Table 3

Labelling efficiency index (LEI) (k-NN) in Scenario 1–3 (%).

N = 6N = 12N = 18N = 24N = 30N = 36N = 42N = 48N = 54N = 60
SCENARIO 1 US 14.2912.7811.3213.2313.3313.0112.7613.4214.0013.91
QBC 10.5411.4513.3713.4013.6013.0413.9615.0115.1715.09
EQBC 10.5411.5213.3913.4013.8013.6814.5315.6815.4315.94
ACR 14.46 16.05 15.12 15.59 15.55 15.25 16.51 15.89 16.84 16.74
SCENARIO 2 US 11.8512.5712.7113.0513.0912.6813.6113.7813.9814.03
QBC 11.6711.4311.5912.0011.7911.9112.7214.0614.2314.25
EQBC 11.6111.5211.8011.6911.4511.6013.4914.1013.5014.35
ACR 13.48 14.85 15.12 16.06 15.86 16.00 15.88 15.74 15.92 15.86
SCENARIO 3 US 11.4811.1212.5213.2713.6213.5013.1813.3013.2413.74
QBC 12.3012.5712.5712.5512.6212.8212.8312.3612.7312.78
EQBC 12.3012.6812.7612.7012.7612.9712.5212.4012.4512.64
ACR 13.26 13.46 13.55 14.25 14.31 14.26 14.23 14.07 14.10 14.38
Table 4

LEI (k-NN) in Scenario 4 (%).

N = 7N = 14N = 21N = 28N = 35N = 42N = 49N = 56N = 63
SCENARIO 4 US 7.1310.8412.8213.0713.5916.50 18.34 17.8818.13
QBC 9.138.7013.1013.0714.3915.2015.7616.8618.13
EQBC 9.1311.7610.3612.4412.4415.8016.8816.9418.13
ACR 10.00 11.76 14.99 15.98 16.16 16.50 18.08 18.05 18.13
Table 5

LEI (SVM) in Scenarios 1–3 (%).

N = 6N = 12N = 18N = 24N = 30N = 36N = 42N = 48N = 54N = 60
SCENARIO 1 US 12.9613.0212.2215.4015.5715.8415.4615.5716.0816.22
QBC 10.9911.9212.2412.3912.3612.3612.8513.9314.6815.22
EQBC 10.9611.8912.2812.3912.9213.0212.9514.0415.0815.78
ACR 18.98 18.48 18.02 18.22 17.73 17.63 18.07 17.88 18.00 18.64
SCENARIO 2 US 12.9613.9312.7913.8614.2515.0915.9715.9915.8815.96
QBC 15.15 16.81 16.6716.2316.5417.1116.85 17.22 17.0017.42
EQBC 13.1315.1915.3015.2514.7515.3816.1715.9716.0916.56
ACR 14.8516.73 17.17 16.73 17.15 17.14 17.03 17.21 17.53 17.46
SCENARIO 3 US 13.8316.0015.52 17.18 17.14 16.79 17.48 16.9817.3917.32
QBC 13.7114.1315.1415.1515.2214.3014.0014.1914.1713.94
EQBC 13.7113.7614.7715.4215.2214.6914.9414.7214.7614.77
ACR 14.29 16.66 16.62 16.7115.95 17.64 16.68 17.60 17.69 18.20
Table 6

LEI (SVM) in Scenario 4 (%).

N = 7N = 14N = 21N = 28N = 35N = 42N = 49N = 56N = 63
SCENARIO 4 US 15.6512.2413.6214.4914.4317.5119.4619.1218.98
QBC 14.5215.6116.26 17.12 16.9017.1618.0818.5218.98
EQBC 16.15 17.14 17.21 16.94 18.14 17.4217.6518.7818.98
ACR 15.6512.2413.6214.4914.43 17.51 19.46 19.12 18.98
Table 7

LEI (RBFNN) in Scenarios 1–3 (%).

N = 6N = 12N = 18N = 24N = 30N = 36N = 42N = 48N = 54N = 60
SCENARIO 1 US 9.0510.369.169.2011.2710.7711.0511.0012.9112.49
QBC 9.6210.2713.3012.6113.0514.1414.3213.6013.9014.33
EQBC 9.6210.2513.6712.5314.8014.4113.3114.1713.5115.19
ACR 12.37 14.68 16.87 18.03 18.08 17.59 17.68 17.57 18.25 18.15
SCENARIO 2 US 14.3914.5614.4015.0315.1813.6813.7315.0315.8315.68
QBC 13.6515.6214.4014.9914.6415.8216.1115.8816.0315.87
EQBC 13.5715.0214.6914.8814.0815.1515.3316.3815.7215.73
ACR 16.86 18.57 17.63 18.12 17.70 17.63 18.49 18.22 18.14 18.79
SCENARIO 3 US 14.7014.6114.5315.2316.9516.7716.3316.4516.9016.71
QBC 12.8113.6813.4113.0412.6613.1512.1112.3513.4413.38
EQBC 12.8112.8513.2612.3912.0613.7613.2714.0314.1213.99
ACR 16.12 18.28 17.07 18.18 17.92 17.43 17.89 17.82 18.20 17.67
Table 8

LEI (RBFNN) in Scenario 4 (%).

N = 7N = 14N = 21N = 28N = 35N = 42N = 49N = 56N = 63
SCENARIO 4 US 8.2310.828.989.2510.2613.2114.1613.8613.50
QBC 9.3510.10 15.40 13.4212.3812.8613.13 14.63 13.50
EQBC 9.3510.8213.13 15.06 11.7612.60 14.59 11.6513.50
ACR 12.35 10.82 13.7014.33 13.80 13.65 13.9913.86 13.50

5. Conclusions

To solve the online drift problem, we redefine a novel AL method, namely AL-ACR, to overcome the imbalance problem of drift information. Instance selection is performed with adaptive rules. Experimental results prove that the proposed method overmatches other references in online working. Furthermore, AL-ACR has obvious advantages in recognition, parameter sensitivity, instance equilibrium and labelling efficiency. It is a favorable choice for the online drift counteraction of E-noses. Future work should be focused on the online mechanism of pool refreshing and online AL implementation in a limited storage resource.
  5 in total

1.  Performance Analysis of ICA in Sensor Array.

Authors:  Xin Cai; Xiang Wang; Zhitao Huang; Fenghua Wang
Journal:  Sensors (Basel)       Date:  2016-05-05       Impact factor: 3.576

2.  Online Active Learning Ensemble Framework for Drifted Data Streams.

Authors:  Jicheng Shan; Hang Zhang; Weike Liu; Qingbao Liu
Journal:  IEEE Trans Neural Netw Learn Syst       Date:  2018-07-02       Impact factor: 10.451

3.  Training and Validating a Portable Electronic Nose for Lung Cancer Screening.

Authors:  Rens van de Goor; Michel van Hooren; Anne-Marie Dingemans; Bernd Kremer; Kenneth Kross
Journal:  J Thorac Oncol       Date:  2018-02-06       Impact factor: 15.609

4.  Metal oxide gas sensor drift compensation using a two-dimensional classifier ensemble.

Authors:  Hang Liu; Renzhi Chu; Zhenan Tang
Journal:  Sensors (Basel)       Date:  2015-04-30       Impact factor: 3.576

5.  Electronic Nose Testing Procedure for the Definition of Minimum Performance Requirements for Environmental Odor Monitoring.

Authors:  Lidia Eusebio; Laura Capelli; Selena Sironi
Journal:  Sensors (Basel)       Date:  2016-09-21       Impact factor: 3.576

  5 in total
  4 in total

1.  Ensemble Learning-Based Approach for Gas Detection Using an Electronic Nose in Robotic Applications.

Authors:  Han Fan; Erik Schaffernicht; Achim J Lilienthal
Journal:  Front Chem       Date:  2022-04-28       Impact factor: 5.545

2.  Active Learning on Dynamic Clustering for Drift Compensation in an Electronic Nose System.

Authors:  Tao Liu; Dongqi Li; Jianjun Chen; Yanbing Chen; Tao Yang; Jianhua Cao
Journal:  Sensors (Basel)       Date:  2019-08-19       Impact factor: 3.576

3.  Sensor Drift Compensation Based on the Improved LSTM and SVM Multi-Class Ensemble Learning Models.

Authors:  Xia Zhao; Pengfei Li; Kaitai Xiao; Xiangning Meng; Lu Han; Chongchong Yu
Journal:  Sensors (Basel)       Date:  2019-09-05       Impact factor: 3.576

Review 4.  Peptides, DNA and MIPs in Gas Sensing. From the Realization of the Sensors to Sample Analysis.

Authors:  Sara Gaggiotti; Flavio Della Pelle; Marcello Mascini; Angelo Cichelli; Dario Compagnone
Journal:  Sensors (Basel)       Date:  2020-08-08       Impact factor: 3.576

  4 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.