| Literature DB >> 33267498 |
Nan Xue1,2,3, Xiong Luo1,2,3, Yang Gao4, Weiping Wang1,2,3, Long Wang1,2,3, Chao Huang1,2,3, Wenbing Zhao5.
Abstract
Kernel adaptive filtering (KAF) is an effective nonlinear learning algorithm, which has been widely used in time series prediction. The traditional KAF is based on the stochastic gradient descent (SGD) method, which has slow convergence speed and low filtering accuracy. Hence, a kernel conjugate gradient (KCG) algorithm has been proposed with low computational complexity, while achieving comparable performance to some KAF algorithms, e.g., the kernel recursive least squares (KRLS). However, the robust learning performance is unsatisfactory, when using KCG. Meanwhile, correntropy as a local similarity measure defined in kernel space, can address large outliers in robust signal processing. On the basis of correntropy, the mixture correntropy is developed, which uses the mixture of two Gaussian functions as a kernel function to further improve the learning performance. Accordingly, this article proposes a novel KCG algorithm, named the kernel mixture correntropy conjugate gradient (KMCCG), with the help of the mixture correntropy criterion (MCC). The proposed algorithm has less computational complexity and can achieve better performance in non-Gaussian noise environments. To further control the growing radial basis function (RBF) network in this algorithm, we also use a simple sparsification criterion based on the angle between elements in the reproducing kernel Hilbert space (RKHS). The prediction simulation results on a synthetic chaotic time series and a real benchmark dataset show that the proposed algorithm can achieve better computational performance. In addition, the proposed algorithm is also successfully applied to the practical tasks of malware prediction in the field of malware analysis. The results demonstrate that our proposed algorithm not only has a short training time, but also can achieve high prediction accuracy.Entities:
Keywords: conjugate gradient; correntropy; kernel adaptive filtering; malware prediction; sparsification criterion
Year: 2019 PMID: 33267498 PMCID: PMC7515314 DOI: 10.3390/e21080785
Source DB: PubMed Journal: Entropy (Basel) ISSN: 1099-4300 Impact factor: 2.524
Computational cost for dictionary update. KLMS (kernel least mean squares); KRLS (kernel recursive least squares); KCG (kernel conjugate gradient); KMCCG (kernel mixture correntropy conjugate gradient).
| Algorithm | Additions | Multiplications | Divisions |
|---|---|---|---|
| KLMS |
|
| 0 |
| KRLS |
|
| 1 |
| KCG |
|
| 3 |
| KMCCG |
|
| 5 |
Figure 1Mackey–Glass time series prediction: learning performance in terms of the testing MSE (mean square error) for QKLMS (quantized kernel least mean squares), QKMC (quantized kernel maximum correntropy), KMMCC (kernel maximum mixture correntropy), and KMCCG (kernel mixture correntropy conjugate gradient) under different noise environments: (a) Gaussian; (b) Bernoulli; (c) sine wave; (d) uniform.
Figure 2Minimum daily temperatures time series prediction: learning performance in terms of the testing MSE for QKLMS, QKMC, KMMCC, and KMCCG.
Figure 3Flowchart for building a malware corpus.
Figure 4Malware API (application programming interface) call time series report.
Figure 5Original malware API call time series.
Figure 6Normalized malware API call time series.
Figure 7Testing MSEs of QKLMS, KMMCC, ANN (artificial neural network), SVM (support vector machine), and KMCCG for malware API call time series prediction.
Computational results of QKLMS, KMMCC, ANN, SVM, and KMCCG in API call time series prediction.
| Algorithm | Time (s) | MSE (dB) |
|---|---|---|
| QKLMS | 39.03 | −12.9139 |
| KMMCC | 52.46 | −15.4081 |
| ANN | 841.41 | −18.1565 |
| SVM | 106.21 | −17.8825 |
| KMCCG | 2.24 | −17.8421 |
Figure 8Testing MSEs of QKLMS, KMMCC, ANN, SVM, and KMCCG for malware API call time series prediction in the noise environment.