| Literature DB >> 35368933 |
Soroush Baseri Saadi1, Nazanin Tataei Sarshar2, Soroush Sadeghi3, Ramin Ranjbarzadeh4, Mersedeh Kooshki Forooshani5, Malika Bendechache4.
Abstract
One of the leading algorithms and architectures in deep learning is Convolution Neural Network (CNN). It represents a unique method for image processing, object detection, and classification. CNN has shown to be an efficient approach in the machine learning and computer vision fields. CNN is composed of several filters accompanied by nonlinear functions and pooling layers. It enforces limitations on the weights and interconnections of the neural network to create a good structure for processing spatial and temporal distributed data. A CNN can restrain the numbering of free parameters of the network through its weight-sharing property. However, the training of CNNs is a challenging approach. Some optimization techniques have been recently employed to optimize CNN's weight and biases such as Ant Colony Optimization, Genetic, Harmony Search, and Simulated Annealing. This paper employs the well-known nature-inspired algorithm called Shuffled Frog-Leaping Algorithm (SFLA) for training a classical CNN structure (LeNet-5), which has not been experienced before. The training method is investigated by employing four different datasets. To verify the study, the results are compared with some of the most famous evolutionary trainers: Whale Optimization Algorithm (WO), Bacteria Swarm Foraging Optimization (BFSO), and Ant Colony Optimization (ACO). The outcomes demonstrate that the SFL technique considerably improves the performance of the original LeNet-5 although using this algorithm slightly increases the training computation time. The results also demonstrate that the suggested algorithm presents high accuracy in classification and approximation in its mechanism.Entities:
Mesh:
Year: 2022 PMID: 35368933 PMCID: PMC8967525 DOI: 10.1155/2022/4703682
Source DB: PubMed Journal: J Healthc Eng ISSN: 2040-2295 Impact factor: 2.682
Figure 1Illustration of LeNet-5 architecture, a convolution neural network. Each plan in the network indicates a feature map [48].
Properties of the layers of the LeNet-5 [48].
| Layer | Size | Num. of feature maps | Num. of parameters | Num. of connections |
|---|---|---|---|---|
| Input | 32 × 32 | … | … | … |
| C1 | 28 × 28 | 6 | 156 | 122304 |
| S2 | 14 × 14 | 6 | 12 | 5880 |
| C3 | 10 × 10 | 16 | 1516 | 151600 |
| S4 | 5 × 5 | 16 | 32 | 2000 |
| C5 | 1 × 1 | 120 | … | 48120 |
| F6 | 1 × 1 | 84 | 10164 | … |
Figure 2Flowchart of SFL algorithm.
Figure 3Datasets sample pictures. (a) OxFord flowers 17. (b) OxFord flowers 102. (c) Caltech/UCSD birds. (d) Caltech 101 airplanes.
Dataset specifications.
| Dataset | Number of categories | Number of images per category | Training sample numbers | Test sample numbers |
|---|---|---|---|---|
| OxFord flowers 17 | 17 | 80 | 50 | 30 |
| OxFord flowers 102 | 102 | 40 to 258 | 20 to 200 | 20 to 58 |
| Caltech/UCSD birds | 200 | 6033 | 4000 | 2033 |
| Caltech 101 airplanes | 101 | 40 to 800 | 20 to 600 | 20 to 200 |
The initial parameters of algorithms.
| Algorithm | Parameter | Value |
|---|---|---|
| SFL | Maximum permitted change in a frog's location | 10 |
| Number of memeplex | 20 | |
| Number of frogs | 30 | |
|
| ||
| ACO | Pheromone update constant ( | 15 |
| Global pheromone decay rate ( | 0.7 | |
| Visibility sensitivity ( | 7 | |
| Population size | 70 | |
| Number of ants | 15 | |
| Maximum number of iterations | 35 | |
| Local pheromone decay rate ( | 0.6 | |
| Initial pheromone ( | 1e-06 | |
| Pheromone sensitivity ( | 1 | |
| Pheromone constant ( | 1 | |
|
| ||
| BFSO | Probability of elimination | 0.1 |
| Spreading percentage % | 0.4 | |
| Population size | 60 | |
| Number of bacteria | 20 | |
| Maximum number of iterations | 35 | |
|
| ||
| WHO | Strategies | Decreasing the value of |
| Whales attacking | Encircling | |
| Max a | 5 | |
| Probability of choosing spiral model |
| |
| Probability of choosing shrinking encircling |
| |
| Population size | 70 | |
| Number of whales | 15 | |
| Maximum number of iterations | 35 | |
Figure 4Error of classification using different datasets and optimization algorithms for LeNet-5. (a) OxFord flowers 17. (b) OxFord flowers 102. (c) Caltech/UCSD birds. (d) Caltech 101 airplanes.
Experimental results for the oxford flowers 17 dataset.
| Technique | MSE (AVE ± STD) | Classification rate (%) | Training time |
|---|---|---|---|
| LeNet 5 | 0.190425 ± 0.031687 | 88 | 14 minutes |
| ACO-LeNet 5 | 0.121689 ± 0.011574 | 90 | 16 minutes |
| BFSO-LeNet 5 | 0.085050 ± 0.034945 | 91 | 25 minutes |
| WO-LeNet 5 | 0.032228 ± 0.039778 | 94 | 18 minutes |
| SFLA-LeNet 5 | 0.009210 ± .039100 | 97 | 23 minutes |
Experimental results for the oxFord flowers 102 dataset.
| Technique | MSE (AVE ± STD) | Classification rate (%) | Training time |
|---|---|---|---|
| LeNet 5 | 0.040320 ± 0.002470 | 90 | 51 minutes |
| ACO-LeNet 5 | 0.024881 ± 0.002472 | 95 | 58 minutes |
| BFSO-LeNet 5 | 0.008026 ± 0.007900 | 93 | 67 minutes |
| WO-LeNet 5 | 0.0229 ± 0.0032 | 91 | 64 minutes |
| SFLA-LeNet 5 | 0.003026 ± 0.001500 | 97 | 65 minutes |
Experimental results for the Caltech/UCSD birds dataset.
| Technique | MSE (AVE ± STD) | Classification rate (%) | Training time |
|---|---|---|---|
| LeNet 5 | 0.0321 ± 0.0045 | 92 | 97 minutes |
| ACO-LeNet 5 | 0.0019 ± 8.4257 | 97 | 101 minutes |
| BFSO-LeNet 5 | 0.0078 ± 8.2189 | 93 | 109 minutes |
| WO-LeNet 5 | 0.0045 ± 8.7654 | 96 | 103 minutes |
| SFLA-LeNet 5 | 0.0021 ± 9.4298 | 98 | 105 minutes |
Experimental results for the Caltech 101 airplanes dataset.
| Technique | MSE (AVE ± STD) | Classification rate (%) | Training time |
|---|---|---|---|
| LeNet 5 | 0.050420 ± 0.003170 | 92 | 49 minutes |
| ACO-LeNet 5 | 0.031841 ± 0.004123 | 94 | 52 minutes |
| BFSO-LeNet 5 | 0.072218 ± 0.079235 | 93 | 57 minutes |
| WO-LeNet 5 | 0.00319 ± 0.0042 | 96 | 53 minutes |
| SFLA-LeNet 5 | 0.00286 ± 0.009700 | 97 | 55 minutes |