| Literature DB >> 30065757 |
Chunhui Bao1, Yifei Pu1, Yi Zhang1.
Abstract
In recent years, the research of artificial neural networks based on fractional calculus has attracted much attention. In this paper, we proposed a fractional-order deep backpropagation (BP) neural network model with L2 regularization. The proposed network was optimized by the fractional gradient descent method with Caputo derivative. We also illustrated the necessary conditions for the convergence of the proposed network. The influence of L2 regularization on the convergence was analyzed with the fractional-order variational method. The experiments have been performed on the MNIST dataset to demonstrate that the proposed network was deterministically convergent and can effectively avoid overfitting.Entities:
Mesh:
Year: 2018 PMID: 30065757 PMCID: PMC6051328 DOI: 10.1155/2018/7361628
Source DB: PubMed Journal: Comput Intell Neurosci
Figure 1The topological structure of the neural networks.
Figure 2The relationship between the fractional order of gradient descent method and the neural network performance.
Performances of the algorithms when v>2.
| Size of training set |
|
| ||
|---|---|---|---|---|
| Train Accuracy | Test Accuracy | Train Accuracy | Test Accuracy | |
| 10000 | 88.65% | 83.52% | 76.31% | 72.66% |
| 20000 | 91.04% | 89.52% | 78.93% | 75.97% |
| 30000 | 93.03% | 90.65% | 82.51% | 80.79% |
| 40000 | 93.20% | 90.53% | 82.47% | 80.61% |
| 50000 | 93.02% | 91.23% | 82.53% | 81.60% |
| 60000 | 93.85% | 91.71% | 87.32% | 86.05% |
Optimal Orders and Highest Accuracies.
| Size of training set | Optimal order of training set | Optimal order of testing set | Highest training accuracy | Highest testing accuracy |
|---|---|---|---|---|
| 10000 | 10/9 | 11/9 | 98.53% | 90.31% |
| 20000 | 10/9 | 10/9 | 98.84% | 92.34% |
| 30000 | 11/9 | 11/9 | 99.05% | 93.50% |
| 40000 | 10/9 | 11/9 | 99.18% | 93.92% |
| 50000 | 1 | 10/9 | 99.20% | 94.56% |
| 60000 | 11/9 | 11/9 | 99.20% | 95.00% |
Performance comparison of different type BP neural networks.
| Size of training set | Integer-order BP neural networks | Fractional-order BP neural networks | Integer-order BP neural networks with | Fractional-order BP neural networks with | ||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Training Accuracy | Testing Accuracy | Training Accuracy | Testing Accuracy | Training Accuracy | Testing Accuracy | Training Accuracy | Testing Accuracy | Improvement relative to IOBP | Improvement relative to FOBP | |
| 10000 | 98.41% | 89.87% | 98.48% | 90.31% | 98.45% | 93.35% | 98.43% | 93.95% | 4.54% | 4.03% |
| 20000 | 98.81% | 92.28% | 98.84% | 92.34% | 98.75% | 95.09% | 98.79% | 95.13% | 3.09% | 3.02% |
| 30000 | 98.95% | 93.38% | 99.05% | 93.50% | 98.92% | 95.15% | 98.88% | 95.62% | 2.40% | 2.27% |
| 40000 | 99.05% | 93.83% | 99.01% | 93.92% | 98.96% | 95.63% | 98.95% | 95.83% | 2.13% | 2.03% |
| 50000 | 99.20% | 94.55% | 99.17% | 94.56% | 99.11% | 96.08% | 99.15% | 96.45% | 2.01% | 2.00% |
| 60000 | 99.17% | 94.87% | 99.20% | 95.00% | 99.13% | 96.51% | 99.17% | 96.70% | 1.93% | 1.79% |
We use the following formula to calculate improvement: improvement of A compared with B = (A-B)÷B.
Figure 3Performance comparison in terms of testing accuracy.
Figure 4Changes of total error E during the training process.
Figure 5Changes of DE during the training process.