Literature DB >> 27077860

Enhancing Electronic Nose Performance Based on a Novel QPSO-KELM Model.

Chao Peng1, Jia Yan2, Shukai Duan3, Lidan Wang4, Pengfei Jia5, Songlin Zhang6.   

Abstract

A novel multi-class classification method for bacteria detection termed quantum-behaved particle swarm optimization-based kernel extreme learning machine (QPSO-KELM) based on an electronic nose (E-nose) technology is proposed in this paper. Time and frequency domain features are extracted from E-nose signals used for detecting four different classes of wounds (uninfected and infected with Staphylococcu aureus, Escherichia coli and Pseudomonas aeruginosa) in this experiment. In addition, KELM is compared with five existing classification methods: Linear discriminant analysis (LDA), quadratic discriminant analysis (QDA), extreme learning machine (ELM), k-nearest neighbor (KNN) and support vector machine (SVM). Meanwhile, three traditional optimization methods including particle swarm optimization algorithm (PSO), genetic algorithm (GA) and grid search algorithm (GS) and four kernel functions (Gaussian kernel, linear kernel, polynomial kernel and wavelet kernel) for KELM are discussed in this experiment. Finally, the QPSO-KELM model is also used to deal with another two experimental E-nose datasets in the previous experiments. The experimental results demonstrate the superiority of QPSO-KELM in various E-nose applications.

Entities:  

Keywords:  electronic nose; feature extraction; kernel extreme learning machine; quantum-behaved particle swarm optimization

Mesh:

Year:  2016        PMID: 27077860      PMCID: PMC4851034          DOI: 10.3390/s16040520

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.576


1. Introduction

An electronic nose (E-nose), combined with artificial intelligence algorithms, is designed for mimicking the mammalian olfactory system to recognize gases and odors. The gas sensor array in an E-nose comprises several non-specific sensors and will generate characteristic patterns when exposed to odorant materials. Patterns of known odorants can be used to construct a database and train a pattern recognition model through quite a few pattern recognition algorithms. In this way, something unknown which can be discriminated by its odor is classified well [1,2,3]. During the past decades, much work has been done to investigate the E-nose technology which has been widely used in a multitude of fields, such as food quality control [4,5,6,7], disease diagnosis [8,9,10,11], environment quality assessment [12,13] and agriculture [14,15,16]. Previous work has proved the effectiveness of detecting bacteria by investigating volatile organic compounds (VOCs) emitted from cultures and swabs taken from patients with infected wounds [17,18,19]. In the pattern recognition, firstly, training data are employed to train the classifier. Then, the performance of this classifier is assessed by using the remaining independent testing samples. The final accuracy can be computed by comparing predicted classes with their true classes. So far, various kinds of classification models have been explored in E-nose applications, which can generally be divided into two categories. One is the linear classifier, such as k-nearest neighbor (KNN) [20,21,22], linear discriminant analysis (LDA) [23,24], partial least squares regression (PLSR) [25] and Bayes classifier [26,27], which is simple and relatively easy to construct but performs poor when it deals with a host of nonlinear problems in E-nose data processing. Another is nonlinear classification models such as the multilayer perceptron (MP) [28,29], radial basis function neural network (RBFNN) [24,30] and decision tree (DT) [31]. The nonlinear classifiers can not only fully approximate the nonlinear relationship of the data, but also show exceedingly strong robustness and fault tolerance. However, they show slow convergence, require too much learning time and are liable to get trapped in local optima. Support vector machine (SVM) is a pretty promising machine learning method that has been widely applied in classification of E-nose data, especially in some complex odor discriminations. It has better results than many classifiers, not only in qualitative and quantitative analysis of E-nose results but also in other applications [32,33,34]. Extreme learning machine (ELM), a fast learning algorithm for single hidden layer feedforward neural networks (SLFNs), first proposed by Huang et al. [35], randomly generates the hidden node parameters and then analytically determines the output weights instead of iterative tuning. Therefore, ELM runs fast, is easy to implement and shows superiority over other classifiers [36]. Nowadays, ELM has been widely used in a range of fields, such as sales forecasting [37], mental tasks [38], face recognition [39] and food quality tracing [40]. However, the classification performance is obviously affected by the algorithm parameters. Meanwhile, the randomly generated input weights and hidden layer biases of ELM can make the algorithm unstable [41]. Kernel Extreme Learning Machine (KELM) is constructed based on ELM combined with kernel functions in this paper considering the above limiting factors. It not only has a good deal of the advantages of ELM, but also can nonlinearly map nonlinear inseparable patterns to a separable high-dimensional feature space, which further improves the accuracy of discriminations. However, due to the existence of kernel functions, KELM is sensitive to the kernel parameters settings. Thus, the quantum-behaved particle swarm optimization (QPSO) is used to optimize the parameters of KELM and in this paper and the QPSO-KELM method is applied to improving the classification accuracy of wound infection detection. The results demonstrate that the proposed method can obtain excellent classification performance in E-nose applications.

2. Materials and Experiments

The datasets used in the paper were obtained by a home-made E-nose, which details can be found in our previous publication [42]. However, to make the paper self-contained, the system structure and experimental setup are briefly repeated here.

2.1. E-Nose System

The sensor array in the research is constructed due to the high sensitivity and quick response of the sensors to the metabolites of three different bacteria. The E-nose system consists of 15 sensors: Fourteen metal oxide gas sensors (TGS800, TGS813, TGS816, TGS822, TGS825, TGS826, TGS2600, TGS2602, TGS2620, WSP2111, MQ135, MQ138, QS-01 and SP3S-AQ2) and one electrochemical sensor (AQ sensor). A 14-bit data acquisition system (DAS) is used as interface between the sensor array and a computer. The DAS converts analog signals from sensor array into digital signals which are stored in the computer for further processing.

2.2. Experimental Setup

Figure 1 shows the schematic diagram of the experimental system. It can be observed that the E-nose system is composed of an E-nose chamber, a data acquisition system (DAS), a pump, a rotor flowmeter, a triple valve, a filter, a glass wild-mouth bottle and a computer. The filter is used to purify the air. The pump is used to convey the VOCs and clean air over the sensor array. The rotor flowmeter is used to control the flow rate during the experiments. The three-way valve is used for switch between VOCs and clean air. The experimental setup has also been mentioned in [33]. The experimental procedure in this paper can be summarized as follows.
Figure 1

Schematic diagram of the experimental system.

Each mouse was put in a big glass bottle with a rubber stopper. Two holes were made in the rubber stopper with two thin glass tubes inserted. One longer glass tube was used as an exit pipe and hung above the wound as close as possible while the shorter one was used as an intake-tube, inserted into the glass a little and was close to the bottleneck. The gases which contained the VOCs of the wound on the mouse outflowed along the longer glass tube and flowed into the sensor chamber. The air flowed into the glass along the shorter glass tube. Each test process comprises three stages: the baseline stage, the response stage and the recovery stage. In the baseline stage, the three-way valve switched on Port 1 and the clean air purified by the filter flowed through the sensor chamber for 3 min. In the response stage, the three-way valve switched on Port 2 and the gases containing the VOCs of the wound flowed through the sensor chamber for 5 min. In the recovery stage, the three-way valve switched on Port 1 again and the clean air flowed through the sensor chamber for 15 min. During the three stages of one test, the DAS always sampled the data and stored them in the computer. After one test and before the next one, for eliminating the influence of the residual odors, the sensor chamber was purged by the clean air for 5 min and in the purging process the DAS did not sample the data. Four groups of mice were tested in the research, including one control group and three groups infected by Staphylococcu aureus, Escherichia coli and Pseudomonas aeruginosa, respectively. Twenty tests for each groups of mice in the same conditions were made, and finally 80 samples for all four groups of mice were collected from the above procedures. Figure 2 illustrates the sensor responses process when they are exposed to four different target wounds, where X-axis is the response time of the sensors and Y-axis is the output voltage of the sensors.
Figure 2

E-nose response to four wounds.

3. Methodology

3.1. KELM

ELM [36,41,43,44,45] is designed as a single hidden layer feed forward network and has been proved that its learning speed is extremely fast. It provides efficient unified solutions to generalized SLFNs, whose hidden nodes can be any piecewise nonlinear function. KELM generalizes ELM from explicit activation to implicit mapping function and produce better generalization in most applications. A brief introduction of KELM is as follows: Suppose there are N training samples (,) where = [x1,x2,…,x] ∈ denotes one sample point in the n-dimensional space and = [t1,t2,…,t] ∈ is the sample class label. The SLFNs and activation function are defined as: where is the i-th sample, L is the number of hidden nodes, and denote the input weights to the hidden layer and the output weight linking the j-th hidden node to the output layer respectively. Meanwhile, b is bias of the j-th hidden node and is the output vector of the input sample . Then, this SLFN can approximate those N samples with zero error, which means that: where is the sample class label vector of the input sample . That is to say, there exist , and b such that: This can be written as: Then, Equation (4) can be also written as matrix form: where is hidden layer output matrix. Then, training such an SLFN is equivalent to finding a least-square solution as follows: where H+ is the Moore-Penrose generalized inverse of the hidden layer output matrix H. Huang et al. suggested adding a positive value 1/C (C is regularization coefficient) to calculate the output weights as follows according to the ridge regression theory: The output function for the SLFN is: where () is the output of the hidden nodes and actually maps the data from input space to the hidden layer feature space H. Thus, substitute Equation (7) into Equation (8), the output function can be defined as follows: We define a kernel function k as: and then a KELM can be constructed using the kernel function exclusively, without having to consider the mapping explicitly. We express this kernel function by Equation (11) for given classes p and q: Let K be a N × N matrix and K = () where , is a matrix composed of inner the product in the feature space: where S is the number of the total classes, N and N are the number of the samples in p-th and q-th classes respectively, is a (N × N) matrix and K is a symmetrical matrix such that . We can define the kernel matrix K = HH from Equation (10) and the output function of KELM can be written as: Some common kernel functions including linear kernel function, polynomial kernel function, Gaussian kernel function, wavelet kernel function are applied. Kernel parameters of the kernel functions, together with regularization coefficient C in Equation (13) will be optimized by QPSO. In this way, the index of the output node with the highest output value is considered as the label of the input data [44].

3.2. QPSO-KELM Model

It is well known that the parameters in algorithms will affect the performances. Therefore, QPSO [46] is used to optimize the value of C in Equation (10) and parameters of the kernel function. The dimension of searching space is corresponding to the number of parameters of KELM with different kernel functions, and the position of each particle represents the parameter values of kernel functions. Because the best generalization performance of KELM can be optimized by QPSO, the testing accuracy can be used as the fitness function of QPSO. The specific steps of QPSO-KELM are described as follows. Normalize all the dataset extracted from the E-nose signals into the range [0,1] and the number of iterations and the population size are set as 30 and 400. Initialize the position and local optimal position of each candidate particle, as well as global best position of the swarm. Calculate each particle’s fitness value according to the fitness function. Update the local optimal positions and global best position. Update the position of each candidate particle in each iteration, which can be calculated by Equation (13). Check the termination criterion. If the maximum number of iterations is not yet reached, return to Step 3 or else go to the Step 6. The best combination of parameters of the kernel function can be acquired, which result in the maximal fitness value. The flowchart of this procedure is illustrated in Figure 3.
Figure 3

Computational procedure of QPSO for optimizing KELM.

4. Results and Discussion

Different features which are able to effectively represent the response of sensors are extracted from the time domain and frequency domain in order to evaluate the effectiveness of the proposed model. The peak value, the integral in the response stage, coefficients of Fourier coefficients (the DC component and first order harmonic component), and approximation coefficients of db1 wavelet of sensor response curve are chosen to be on behalf of the characteristics of E-nose signals from two transform domains [47,48,49,50]. Then, leave-one-out cross validation (LOO-CV) method is employed to evaluate the performances of different methods in this experiment for making full use of the data set. Another five classification models, namely ELM, SVM, KNN, LDA and quadratic discriminant analysis (QDA), are applied for comparison with KELM. ELM is an algorithm for single-hidden layer feed forward networks training that leads to fast networking requiring low human supervision. The main idea in ELM is that the network hidden layer parameters need not to be learned, but can be randomly assigned. The only parameter is the number of hidden nodes in the hidden layer of SLFN, which is normally obtained by a trial and error method. Thus, the input weights are within (−1, 1) and the hidden layer biases are within (0, 1). 100 experiments were carried out according to the number of hidden nodes in the hidden layer from 1 to 100. Because the input weights and the hidden layer biases were chosen randomly, this experiment was repeated for 100 times. The best performance of all results will be regarded as the final classification results of ELM. For SVM, LIBSVM is employed in this paper, which is devolved by Chang and Lin [51]. KNN requires two parameters to tune: The number of neighbor k and the distance metric. In this work, the values of k vary from 1 to 20, and several distance metrics which are used are Euclidean distance, cityblock distance, cosine distance and correlation distance. The best classification accuracy of different values of k and distance metrics will be regarded as the final results of the KNN. Table 1, Table 2, Table 3 and Table 4 list the classification results of the four feature extraction techniques and five classification models. The kernel function of KELM is set to Gaussian kernel. The bold type numbers in diagonal indicate the number of samples classified correctly, while others indicate the number of samples misclassified.
Table 1

Classification results of peak value.

ClassPredicted as *
KELMELMSVMLDAKNNQDA
123412341234123412341234
1173001630116400146001640015302
2119004142051500911002171031610
3111620014600173101630315204133
4102170121700416005151121620315
Total86.25%76.25%80.00%70.00%80.00%73.75%

* 1, No-infection; 2, Staphylococcu aureus; 3, Escherichia coli; 4, Pseudomonas aeruginosa, similarly hereinafter.

Table 2

Classification results of integral value.

ClassPredicted as *
KELMELMSVMLDAKNNQDA
123412341234123412341234
1173001541015500137001730016310
2317006131041600812002180021440
3101810117200173101720315200173
4000200031700218004160161300416
Total90.00%77.50%82.50%72.50%78.75%78.75%
Table 3

Classification results of Fourier coefficients.

ClassPredicted as *
KELMELMSVMLDAKNNQDA
123412341234123412341234
1191001730017300164001730018200
2218002180021800317001190001550
3001820216211162001640118100191
4002180041600119004161111700515
Total91.25%83.75%87.50%81.25%88.75%83.75%
Table 4

Classification results of wavelet coefficients.

ClassPredicted as *
KELMELMSVMLDAKNNQDA
123412341234123412341234
1200001541017300173001730020000
2218002180031700515003170011720
3101810018200182101810018200173
4000200121700119002180031700614
Total95.00%85.00%88.75%85.00%86.25%85.00%
It can be observed from the above four tables that the classification accuracy of the four wounds is influenced both by different features and classification models. In general, features extracted from frequency domain can achieve better results, while features extracted from time domain do worse. It can be also seen that the classification effect of wavelet coefficients feature works best no matter what kinds of classifier are used, while peak value feature is just performs worst. QPSO-KELM always performs better than other four classifiers regardless of what kinds of features are used. SVM is invariably performs better than rest three classifiers as well. For wounds uninfected, the best performance is achieved when the wavelet feature is put into the QPSO-KELM model, where there is no sample misclassified; for wounds infected with Staphylococcu aureus, QPSO-KELM performs best when the peak value is used as the feature, in which there is only one sample misclassified; for wounds infected with Escherichia coli, the highest classification accuracy is achieved by QDA with the feature of Fourier coefficients; for wounds infected with Pseudomonas aeruginosa, QPSO-KELM achieves best when features are integral value and wavelet coefficients. Figure 4 and Figure 5 show the variation of the classification rate with the number of hidden nodes in the hidden layer of ELM and the k value of KNN for the priority to classification of wavelet coefficients feature. Figure 4 shows only the classification results of one of the 100 repeated experiments to display the change process with the number of hidden nodes in the hidden layer varying from 1 to 100. It can be clearly seen that the classification rate gradually improves with the number of hidden nodes from 1 to 34 and from 79 to 96 from the Figure 4, while the classification rate gradually declines with the number of hidden nodes from 55 to 79. Moreover, ELM can achieve the best classification accuracy of 85% when is the number of hidden nodes are 45, 51 and 55.
Figure 4

The performance of ELM according to the number of hidden nodes from 1 to 100.

Figure 5

The performance of KNN with different k values and distance metrics.

Figure 5 manifests that the classification rate gradually declines as the k value increases on the whole. For different distance metrics, the cityblock distance performs worst except k = 8, 10, 20. Meanwhile, the cosine distance can achieve the best classification accuracy of 86.25% at the start stage and performs best as well at the last stage. Another three traditional optimization methods are also investigated and used to devaluate the effectiveness of the proposed model when wavelet coefficients are used as features. PSO [52], Genetic algorithm (GA) [53,54] and Grid search algorithm (GS) are employed to optimize parameters of KELM. For GA and PSO, the maximum number of iterations and the population size are also 400 and 30, respectively, which is the same as those of QPSO. For GS, the ranges of the model parameters are set according to [44]. The range of the cost parameter C and the kernel parameter of the Gaussian kernel function are both [2-25,225], and the step length is set as 20.5. Their classification performances are shown in Table 5.
Table 5

Comparison with different optimization methods for KELM.

ClassPredicted as *
QPSOPSOGAGS
1234123412341234
120000182001820017102
221800317003170021800
310181001910019110172
400020003170041610217
Total95.00%88.75%87.50%86.25%
It is obviously that the QPSO-KELM model obtains 95% classification rate, while other traditional methods perform worse than the proposed model from Table 5, especially that it is all predicted correctly for wounds uninfected and wounds infected with Pseudomonas aeruginosa. QPSO-KELM, PSO-KELM, GA-KELM and GS-KELM can only achieve 88.75%, 87.5% and 86.25% classification rates, respectively. It is well known that the choice of kernel function plays a crucial role in recognition and generalization capability. Thus, in order to further explore the effects of different kernel functions on the QPSO-KELM model, the effects of four kinds of common kernel functions combined with wavelet features are investigated in this experiment. Their classification performances of different kernel functions are shown in Table 6.
Table 6

Classification results of four kernel functions used in the QPSO-KELM model.

ClassPredicted as *
GaussianLinearPolynomialWavelet
1234123412341234
120000182001820019100
221800218002180021800
310181001731018110181
400020005150011900119
Total95.00%85.00%91.25%92.50%
It can be clearly concluded that the QPSO-KELM model with Gaussian kernel function performs best from Table 6, while the linear kernel function achieves the worst accuracy. Meanwhile, the performance of the polynomial kernel function is close to that of wavelet kernel function, which achieves 91.25% and 92.50% respectively. It means that the proposed model performs best in all of the above methods. We also use the proposed model to deal with another two experimental E-nose datasets: (1) dataset of an E-nose which recognizes seven bacteria: Pseudomonas aeruginosa, Escherichia coli, Acinetobacter baumannii, Staphylococcu aureus, Staphylococcus epidermidis, Klebsiella pneumoniae and Streptococcus pyogenes. The classification results of various classification models based on steady-state signals of sensors are shown in Table 7. More details concerning the experiment can be found in [55]; (2) dataset of an E-nose which detects six indoor air contaminants including formaldehyde (HCHO), benzene (C6H6), toluene (C7H8), carbon monoxide (CO), ammonia (NH3) and nitrogen dioxide (NO2) and classification results are also shown in Table 8. More details include dataset generation regarding the experiment can be found in [56].
Table 7

Accuracy results of various feature extraction techniques and classification models for datasets in [55].

ClassAccuracy Rate (%)
KELMSVMELMKNNLDAQDA
Pseudomonas aeruginosa100.00100.0080.00100.00100.00100.00
Escherichia coli100.00100.00100.00100.00100.0090.00
Acinetobacter baumannii100.00100.0090.0090.0090.00100.00
Staphylococcu aureus100.0090.0090.0080.0090.0090.00
Staphylococcus epidermidis100.00100.00100.00100.00100.00100.00
Klebsiella pneumoniae100.0090.00100.0080.0080.0070.00
Streptococcus pyogenes100.0080.0090.0060.0060.0060.00
Average100.0094.2992.8687.1488.5787.14
Table 8

Accuracy results of various feature extraction techniques and classification models for datasets in [56].

ClassAccuracy Rate (%)
KELMSVMELMKNNLDAQDA
HCHO94.2394.2392.3190.3894.2363.46
C6H690.9187.8872.7375.7657.5887.88
C7H8100.00100.00100.00100.00100.00100.00
CO100.0091.67100.0091.6783.33100.00
NH370.0060.0070.0070.0080.0080.00
NO2100.00100.0066.6766.6783.3383.33
Average92.5288.9683.6282.4183.0885.78
It can be clearly concluded that the proposed QPSO-KELM model achieves the best classification accuracy among all of the above classification models for different datasets. The KELM achieves the best recognition performance of 100% for the dataset in [55] and can also obtain the best recognition accuracy except the recognition rate 70% of NH3 for the dataset [56]. It demonstrates that the QPSO-KELM approach has outstanding generalized performance with other datasets, which efficacy does not depend on a particular dataset.

5. Conclusions

In this paper, a new methodology based on the QPSO-KELM model has been presented to enhance the performance of an E-nose for wound infection detection. Four kinds of features extracted from the time and frequency domains have been developed to demonstrate the effectiveness of this classification model for four different classes of wounds. It first introduces the kernel method based on extreme learning machine into the E-nose application of this paper, which provides a new idea for signal processing of E-nose data. Moreover, this paper also provides a good solution for the optimization of kernel function parameters by QPSO, which is a contraction mapping algorithm that outperforms ordinary optimization algorithms in the rate of convergence and convergence ability. Experimental tests have been carried out to verify that the proposed QPSO-KELM model can lead to a higher accuracy rate and manifest that the QPSO-KELM model can obviously enhance E-nose performance in various applications. The model in this study also provides an efficient approach in applications related to classification or prediction, not only in E-nose applications, but also in other uses.
  18 in total

1.  Electronic nose analysis of exhaled breath to diagnose ventilator-associated pneumonia.

Authors:  R M Schnabel; M L L Boumans; A Smolinska; E E Stobberingh; R Kaufmann; P M H J Roekaerts; D C J J Bergmans
Journal:  Respir Med       Date:  2015-09-30       Impact factor: 3.415

Review 2.  Exhaled breath analysis by electronic nose in respiratory diseases.

Authors:  Simone Scarlata; Giorgio Pennazza; Marco Santonico; Claudio Pedone; Raffaele Antonelli Incalzi
Journal:  Expert Rev Mol Diagn       Date:  2015-05-11       Impact factor: 5.225

3.  Classification of mental tasks from EEG signals using extreme learning machine.

Authors:  Nan-Ying Liang; Paramasivan Saratchandran; Guang-Bin Huang; Narasimhan Sundararajan
Journal:  Int J Neural Syst       Date:  2006-02       Impact factor: 5.866

4.  Detection of potato brown rot and ring rot by electronic nose: from laboratory to real scale.

Authors:  E Biondi; S Blasioli; A Galeone; F Spinelli; A Cellini; C Lucchese; I Braschi
Journal:  Talanta       Date:  2014-05-02       Impact factor: 6.057

5.  Quality control of Lonicera japonica stored for different months by electronic nose.

Authors:  Yin Xiong; Xiaohe Xiao; Xiaoyun Yang; Dan Yan; Congen Zhang; Huiqin Zou; Hui Lin; Lian Peng; Xiao Xiao; Yonghong Yan
Journal:  J Pharm Biomed Anal       Date:  2013-12-25       Impact factor: 3.935

6.  Development and application of a new electronic nose instrument for the detection of colorectal cancer.

Authors:  E Westenbrink; R P Arasaradnam; N O'Connell; C Bailey; C Nwokolo; K D Bardhan; J A Covington
Journal:  Biosens Bioelectron       Date:  2014-10-23       Impact factor: 10.618

Review 7.  Diverse applications of electronic-nose technologies in agriculture and forestry.

Authors:  Alphus D Wilson
Journal:  Sensors (Basel)       Date:  2013-02-08       Impact factor: 3.576

8.  Applications and advances in electronic-nose technologies.

Authors:  Alphus D Wilson; Manuela Baietto
Journal:  Sensors (Basel)       Date:  2009-06-29       Impact factor: 3.576

Review 9.  Electronic noses for environmental monitoring applications.

Authors:  Laura Capelli; Selena Sironi; Renato Del Rosso
Journal:  Sensors (Basel)       Date:  2014-10-24       Impact factor: 3.576

Review 10.  Electronic Nose Feature Extraction Methods: A Review.

Authors:  Jia Yan; Xiuzhen Guo; Shukai Duan; Pengfei Jia; Lidan Wang; Chao Peng; Songlin Zhang
Journal:  Sensors (Basel)       Date:  2015-11-02       Impact factor: 3.576

View more
  7 in total

1.  Test Case Prioritization, Selection, and Reduction Using Improved Quantum-Behaved Particle Swarm Optimization.

Authors:  Anu Bajaj; Ajith Abraham; Saroj Ratnoo; Lubna Abdelkareim Gabralla
Journal:  Sensors (Basel)       Date:  2022-06-09       Impact factor: 3.847

2.  Radar HRRP Target Recognition Based on Stacked Autoencoder and Extreme Learning Machine.

Authors:  Feixiang Zhao; Yongxiang Liu; Kai Huo; Shuanghui Zhang; Zhongshuai Zhang
Journal:  Sensors (Basel)       Date:  2018-01-10       Impact factor: 3.576

3.  A Novel Extreme Learning Machine Classification Model for e-Nose Application Based on the Multiple Kernel Approach.

Authors:  Yulin Jian; Daoyu Huang; Jia Yan; Kun Lu; Ying Huang; Tailai Wen; Tanyue Zeng; Shijie Zhong; Qilong Xie
Journal:  Sensors (Basel)       Date:  2017-06-19       Impact factor: 3.576

4.  Towards Intelligent Interpretation of Low Strain Pile Integrity Testing Results Using Machine Learning Techniques.

Authors:  De-Mi Cui; Weizhong Yan; Xiao-Quan Wang; Lie-Min Lu
Journal:  Sensors (Basel)       Date:  2017-10-25       Impact factor: 3.576

5.  Feature Extraction of Electronic Nose Signals Using QPSO-Based Multiple KFDA Signal Processing.

Authors:  Tailai Wen; Jia Yan; Daoyu Huang; Kun Lu; Changjian Deng; Tanyue Zeng; Song Yu; Zhiyi He
Journal:  Sensors (Basel)       Date:  2018-01-29       Impact factor: 3.576

6.  Development of Organic-Inorganic Hybrid Optical Gas Sensors for the Non-Invasive Monitoring of Pathogenic Bacteria.

Authors:  Sumana Kladsomboon; Chadinee Thippakorn; Thara Seesaard
Journal:  Sensors (Basel)       Date:  2018-09-21       Impact factor: 3.576

7.  Cross-Category Tea Polyphenols Evaluation Model Based on Feature Fusion of Electronic Nose and Hyperspectral Imagery.

Authors:  Baohua Yang; Lin Qi; Mengxuan Wang; Saddam Hussain; Huabin Wang; Bing Wang; Jingming Ning
Journal:  Sensors (Basel)       Date:  2019-12-20       Impact factor: 3.576

  7 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.