| Literature DB >> 32941468 |
Shengbing Ren1, Fa Liu1, Weijia Zhou1, Xian Feng1, Chaudry Naeem Siddique1.
Abstract
The deep multiple kernel Learning (DMKL) method has attracted wide attention due to its better classification performance than shallow multiple kernel learning. However, the existing DMKL methods are hard to find suitable global model parameters to improve classification accuracy in numerous datasets and do not take into account inter-class correlation and intra-class diversity. In this paper, we present a group-based local adaptive deep multiple kernel learning (GLDMKL) method with lp norm. Our GLDMKL method can divide samples into multiple groups according to the multiple kernel k-means clustering algorithm. The learning process in each well-grouped local space is exactly adaptive deep multiple kernel learning. And our structure is adaptive, so there is no fixed number of layers. The learning model in each group is trained independently, so the number of layers of the learning model maybe different. In each local space, adapting the model by optimizing the SVM model parameter α and the local kernel weight β in turn and changing the proportion of the base kernel of the combined kernel in each layer by the local kernel weight, and the local kernel weight is constrained by the lp norm to avoid the sparsity of basic kernel. The hyperparameters of the kernel are optimized by the grid search method. Experiments on UCI and Caltech 256 datasets demonstrate that the proposed method is more accurate in classification accuracy than other deep multiple kernel learning methods, especially for datasets with relatively complex data.Entities:
Mesh:
Year: 2020 PMID: 32941468 PMCID: PMC7498035 DOI: 10.1371/journal.pone.0238535
Source DB: PubMed Journal: PLoS One ISSN: 1932-6203 Impact factor: 3.240
Fig 1Comparison of two multiple kernel learning methods.
Fig 2Multiple kernel learning.
Fig 3GLDMKL architecture.
Fig 4Multiple kernel k-means clustering.
Fig 5The training and testing process in GLDMKL.
Kernel parameters setting.
| Kernel | Equation | Parameters |
|---|---|---|
| Polynomial | ||
| Laplacian |
| |
| Tanh | ||
| RBF |
| |
| Arc-cosine [ |
|
|
Five classical comparison methods.
| Methods | Detail | Year |
|---|---|---|
| 2LMKL | Zhuang proposed a two-layer multiple kernel learning algorithm [ | In 2011 |
| DMKL | Deep multiple kernel learning algorithm proposed by strobl [ | In 2013 |
| MLMKL | Multi-layer multiple kernel learning algorithm for backpropagation proposed by Rebai [ | In 2016 |
| SA-DMKL | Adaptive deep multiple kernel learning algorithm proposed [ | In 2019 |
| DWS-MKL | Depth-width-scaling multiple kernel learning algorithm proposed [ | In 2020 |
Selected datasets in UCI.
| Datasets | Dimensions | Samples |
|---|---|---|
| Liver | 6 | 345 |
| Breast | 9 | 286 |
| Sonar | 60 | 208 |
| Australian | 14 | 690 |
| German | 20 | 1000 |
| Monk | 6 | 432 |
Comparison of best classification performance(%).
| Datasets | Algorithms | |||||
|---|---|---|---|---|---|---|
| 2LMKL | DMKL | MLMKL | SA-DMKL | DWS-MKL | GLDMKL | |
| Liver | 63.43 | 69.01 | 71.80 | 75.65 | 74.85 | |
| Breast | 96.53 | 96.59 | 97.21 | 91.92 | 97.08 | |
| Sonar | 83.75 | 83.94 | 83.84 | 89.42 | 84.79 | |
| Australian | 82.11 | 84.40 | 85.42 | 82.03 | 85.51 | |
| German | 72.22 | 72.02 | 75.06 | 78.50 | 73.60 | |
| Monk | 96.26 | 96.62 | 96.89 | 97.55 | 99.07 | |
Fig 6Performance comparison(%) of different groups.
Comparison of classification performance(%) at each layer.
| Dataset | Algorithms | ||||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| DMKL | MLMKL | SA-DMKL | DWS-MKL | GLDMKL | |||||||||
| 1-layer | 2-layer | 3-layer | 1-layer | 2-layer | 3-layer | *-layer | 1-layer | 2-layer | 3-layer | 1-layer | 2-layer | 3-layer | |
| Liver | 69.53 | 69.53 | 69.01 | 69.53 | 70.17 | 71.80 | 75.65 | 74.85 | 72.51 | 72.51 | 74.00 | 76.92 | |
| Breast | 94.92 | 96.56 | 96.59 | 96.89 | 96.83 | 97.21 | 91.92 | 96.78 | 97.08 | 97.08 | 92.59 | 72.73 | |
| Sonar | 82.98 | 84.13 | 83.94 | 83.94 | 84.23 | 83.84 | 89.42 | 84.76 | 84.76 | 84.76 | 74.04 | ||
| Australian | 83.56 | 84.26 | 84.40 | 85.13 | 85.56 | 85.42 | 82.03 | 84.35 | 84.64 | 85.51 | |||
| German | 70.56 | 71.48 | 72.02 | 73.80 | 74.56 | 75.06 | 78.50 | 72.00 | 72.00 | 73.60 | 74.25 | 79.01 | |
| Monk | 96.25 | 96.52 | 96.62 | 96.89 | 96.62 | 96.89 | 97.55 | 99.07 | 99.07 | 99.07 | 99.54 | ||
Classification performance(%) at each layer where the bowling-ball class is positive.
| Layers | Groups | ||||
|---|---|---|---|---|---|
| 1 group | 2 groups | 5 groups | 7 groups | 10 groups | |
| 1-layer | 87.22 | 86.55 | 80.00 | ||
| 2-layer | 87.51 | 86.55 | 86.97 | 61.76 | |
| 3-layer | 87.01 | 86.63 | 86.97 | 85.29 | |
| 4-layer | 89.05 | 63.73 | |||
| 5-layer | 89.79 | 75.49 | |||
Classification performance(%) at each layer where the sunflower-101 class is positive.
| Layers | Groups | ||||
|---|---|---|---|---|---|
| 1 group | 2 groups | 5 groups | 7 groups | 10 groups | |
| 1-layer | 87.57 | 87.11 | 88.57 | ||
| 2-layer | 85.56 | 81.48 | 88.57 | ||
| 3-layer | 85.56 | 85.59 | 84.29 | ||
| 4-layer | 85.56 | 85.59 | |||
| 5-layer | 85.59 | ||||
Classification performance(%) at each layer where the car-tire class is positive.
| Layers | Groups | ||||
|---|---|---|---|---|---|
| 1 group | 2 groups | 5 groups | 7 groups | 10 groups | |
| 1-layer | 89.01 | 87.70 | 87.88 | 88.49 | |
| 2-layer | 87.01 | 80.80 | 82.00 | 88.49 | 71.14 |
| 3-layer | 86.52 | 80.80 | 82.00 | 88.49 | 71.14 |
| 4-layer | |||||
| 5-layer | |||||
Classification performance(%) at each layer where the desk-globe class is positive.
| Layers | Groups | ||||
|---|---|---|---|---|---|
| 1 group | 2 groups | 5 groups | 7 groups | 10 groups | |
| 1-layer | 89.36 | 70.00 | 88.63 | 85.63 | 84.12 |
| 2-layer | 55.81 | 71.63 | 89.32 | 85.63 | 84.12 |
| 3-layer | 83.97 | 71.63 | 89.32 | 85.00 | 84.05 |
| 4-layer | 89.32 | 85.00 | 84.05 | ||
| 5-layer | 85.43 | 86.43 | |||
Classification performance(%) at each layer where the roulette-wheel class is positive.
| Layers | Groups | ||||
|---|---|---|---|---|---|
| 1 group | 2 groups | 5 groups | 7 groups | 10 groups | |
| 1-layer | 88.52 | 79.50 | 80.00 | 88.86 | |
| 2-layer | 89.37 | 88.02 | 88.13 | 72.36 | 88.86 |
| 3-layer | 89.37 | 88.13 | 72.36 | 80.68 | |
| 4-layer | 89.37 | 83.30 | |||
| 5-layer | |||||