Literature DB >> 35440609

Hybridized sine cosine algorithm with convolutional neural networks dropout regularization application.

Nebojsa Bacanin1, Miodrag Zivkovic2, Fadi Al-Turjman3, K Venkatachalam4, Pavel Trojovský4,5, Ivana Strumberger2, Timea Bezdan2.   

Abstract

Deep learning has recently been utilized with great success in a large number of diverse application domains, such as visual and face recognition, natural language processing, speech recognition, and handwriting identification. Convolutional neural networks, that belong to the deep learning models, are a subtype of artificial neural networks, which are inspired by the complex structure of the human brain and are often used for image classification tasks. One of the biggest challenges in all deep neural networks is the overfitting issue, which happens when the model performs well on the training data, but fails to make accurate predictions for the new data that is fed into the model. Several regularization methods have been introduced to prevent the overfitting problem. In the research presented in this manuscript, the overfitting challenge was tackled by selecting a proper value for the regularization parameter dropout by utilizing a swarm intelligence approach. Notwithstanding that the swarm algorithms have already been successfully applied to this domain, according to the available literature survey, their potential is still not fully investigated. Finding the optimal value of dropout is a challenging and time-consuming task if it is performed manually. Therefore, this research proposes an automated framework based on the hybridized sine cosine algorithm for tackling this major deep learning issue. The first experiment was conducted over four benchmark datasets: MNIST, CIFAR10, Semeion, and UPS, while the second experiment was performed on the brain tumor magnetic resonance imaging classification task. The obtained experimental results are compared to those generated by several similar approaches. The overall experimental results indicate that the proposed method outperforms other state-of-the-art methods included in the comparative analysis in terms of classification error and accuracy.
© 2022. The Author(s).

Entities:  

Mesh:

Year:  2022        PMID: 35440609      PMCID: PMC9016213          DOI: 10.1038/s41598-022-09744-2

Source DB:  PubMed          Journal:  Sci Rep        ISSN: 2045-2322            Impact factor:   4.996


Introduction

Artificial intelligence (AI) has the goal of creating human-level artificial intelligence (HLAI)[1] and currently, the leading AI subdomain is machine learning, more precisely deep learning. This type of AI is not focused on representing human behavior as the main goal suggests, but rather focuses on providing practically usable results. In this way, deep learning has achieved the levels of human performance regarding specified tasks, however, the HLAI has not yet been achieved[2]. It particularly excels in the tasks of computer vision, natural language processing (NLP), and speech recognition. For the program of this sort to successfully accomplish tasks such as mentioned, its architecture should imitate the behavior of the human nervous system[3-5]. As an attempt to create such architecture, convolutional neural networks (CNNs) have been created[6]. The CNNs are inspired by the animal visual cortex and consist of several layers, where each layer, except the first input layer, takes the output of the previous one and sends it to the next one. This behavior forms the basics of the CNN model, as the input becomes more filtered with each layer. This provides more detailed outputs after each layer, while reducing the input complexity to an easier form to process, and at the same time without losing any of the critical feature data. The classic example of this process is with the forming of edges on the first layer, and later on respectively with each layer, the corners and sets of edges, parts of objects and sets of corners and contours, and on the final layer recognition of full objects based on the sets of the previous forming parts is achieved. However, deep learning models, especially deep neural networks, such are CNNs, suffer from some shortcomings. Two of the most important drawbacks of these models are hyperparameters’ tuning and the overfitting issue[7]. For each specific task, a CNN with certain architecture should be generated, because CNN architecture that performs well on every problem does not exist. The architecture is defined with the number and types of layers, the number of neurons in each layer, the loss function learning rate, type of activation function, etc. These components, which are not trainable, are known as hyperparameters’ and it is very challenging to find their proper values for a specific task. This issue is known in the literature as hyperparameters’ tuning[8]. The major source of the overfitting issue is that during the training, the model’s weight and biases become well adjusted for a limited amount of training data, which makes the model inefficient in making predictions for previously unseen observations (testing data). In other words, the model will not be able to generalize well. This issue can be also viewed from the perspective of bias variance trade-off[9]. For solving the issue of overfitting several regularization techniques have been proposed[7]. However, one of the most efficient methods, especially for complex structures such are CNNs, is dropout[10]. The basic idea behind the dropout is removing random units (neurons) from the CNNs along with all their input and output connections. The dropout is controlled by the parameter known as the dropout probability () that determines the percentage of units that are discarded from the model. However, the dropout probability is another CNNs hyperparameter that should be tuned for each particular problem (dataset). One practical way to address this issue is to use an automated framework that will generate optimal or sub-optimal dropout probability value for a given CNN’s structure and the dataset, instead of finding this value manually by performing “trial and error”. A Group of algorithms that can be very efficient in executing such tasks are metaheuristics. In this manuscript, an automated framework for determining a proper dp value for CNN’s dropout layer, based on the swarm intelligence approaches is presented. From the literature examination, it can be noticed that the first framework for dropout regularization by swarm intelligence was proposed by de Rosa et. al. in 2017[7], where the authors implemented firefly algorithm (FA), cuckoo search (CS), bat algorithm (BA), particle swarm optimization (PSO) for tackling this issue. However, it also can be noticed the lack of swarm intelligence applications for this task and it can be concluded that the swarm intelligence potential in determining optimal (sub-optimal) dropout probability ratio in CNNs was not investigated enough. The research proposed in this manuscript represents an extension of the investigation shown in[11], where the proposed swarm intelligence-based framework for dropout regularization was evaluated for only MNIST and CIFAR-10 datasets. In this manuscript, automated dropout regularization CNNs framework, based on the hybridized sine cosine algorithm (SCA) with the well-known firefly algorithm (FA) swarm metaheuristics, was brought forward and it was first validated against MNIST, CIFAR-10, UPS, and Semeion datasets. In the second experiment, the proposed algorithm has been validated on magnetic resonance images (MRI) classification task. The SCA is a relatively efficient algorithm proposed in 2016 by Mirjalili[12], and since it in its basic implementation exhibits some deficiencies, an improved, hybrid SCA version was devised for the purpose of this research. Before applying hybrid SCA for dropout regularization, simulations with CEC2019 benchmark instances were conducted and these results are reported as well in the manuscript. Also, to further investigate the performance of other swarm intelligence algorithms for this problem, besides basic SCA and proposed hybrid SCA, recent whale optimization (WOA) swarm algorithm is implemented and tested for this problem as well. In the very introduction of this manuscript, some terminology ambiguous should be uncovered. According to some sources, the SCA belongs to the group of math-inspired population-based metaheuristics[13], while some other sources put the SCA in the group of swarm intelligence[14]. In this manuscript, since the hybridization with other swarm algorithm was performed, the SCA is categorized as a swarm intelligence approach. Taking into consideration basic research assumptions, the motivation of the proposed investigation is to further improve CNN’s classifications performance and to avoid overfitting by establishing better dropout regularization than other methods which results are reported in the literature by utilizing novel SCA metaheuristics. In line with the common practice in computer science when devising and testing the new or modified algorithm, the novel metaheuristics has first been validated on a set of challenging CEC2019 benchmark functions. Afterwards, the proposed algorithm was applied on the problem of dropout probability estimation and tested on four benchmark datasets. Finally, the algorithm has been validated on a practical MRI classification task. Therefore, according to empirical results presented in “CEC2019 benchmark simulations” and “CNN dropout regularization simulations” Sections , the contribution of the proposed research is twofold: classification accuracy of CNN model used in simulations is enhanced and novel state-of-the-art SCA metaheuristics was devised. Remain sections of the proposed manuscript are structured as follows: “Background and related work” Section gives a brief theoretical background of CNNs and dropout regularization along with relevant literature review from the swarm intelligence domain. The goal of “Proposed method” Section is to provide a basic description of the original SCA and its drawbacks, as well as to introduce insight into the proposed hybrid SCA metaheuristics. “CEC2019 benchmark simulations” Section provides experimental results and comparative analysis of the proposed method for standard CEC2019 benchmarks following with “CNN dropout regularization simulations” Section, where the results obtained from simulations for CNN’s dropout regularization are reported. Finally, “Conclusion” Section concludes this paper and shows directions of potential further research from this area.

Background and related work

Human beings are not able to process any of information absorbed by labeling, tagging, and putting it into tables. This creates a limitation for the accurate representation of the information obtained in the computational form. It is inefficient and too complex to process for an individual and to translate the obtained information from the photograph into words and in a way that a program can process them. For this reason, the CNN technology has been widely applied for use in visual tasks[15] and nowadays it is the most commonly utilized deep learning model[6]. The recent advancements employ facial recognition[16-19], document analysis[20,21], medical images classification and diagnostics[22-24], and an important task of analyzing climate change and extreme weather conditions[25,26] among many other. The CNN, besides the input layer, consists of three basic types of layers: convolution, pooling, and the fully-connected (dense) layers. The convolution layers filter the data by applying the convolution operation and the features are extracted by the filters of sizes always smaller than input. The most common filter (kernel) sizes are 3 3, 5 5, and 7 7. When kernels are applied to the input, feature maps are generated. Mathematical representation of the convolution operation on the input vector is represented as follows:where stands for the value of the output feature of the k-th feature map at location i, j. Representation of the input is given as x at the location i, j, w denotes the filters, and bias is b. Following up the convolution operation is the activation given as:where the is the non-linear function using the output. Pooling layers can be global or local and the two most applied types are max and average pooling layers. The pooling function is used to reduce the resolution: Fully connected layers in CNNs perform the same operations as in classic ANNs. Typical CNN may consist of several dense layers, where the last layer performs classification by using the softmax for multi-class classification and the sigmoid or tanh activation function for binary classification tasks. Despite the diverse application that CNN technology offers, it is not without shortcomings. As previously noted in “Introduction” Section, CNNs suffer from overfitting and require ways of avoiding such scenarios[10,27]. The most common methods used to address overfitting are[7]: model simplification, early stopping, data augmentation, and regularization. Many regularization techniques have been proposed, e.g. L1[28] and L2 regularization (weight-decay)[28] and dropout[10]. To drop a unit from a layer means that it is removed with all of its connections. The neurons to be dropped are selected randomly and temporarily removed from the training process. The absence of these neurons is believed to result in a network with better generalization because it becomes less sensitive to the weight of those neurons. The basis is to exclude randomly selected hidden individuals during the training phase. The goal is to force neighboring neurons to take over the workload from missing neurons which leads to an increase in independent internal representations. This process is only performed ahead of the classification layer upon the last fully-connected layers. The feed-forward operation is performed as the following equation:where the symbols represent the following terms: l is the l-th hidden layer of the network, z and y are input and output vectors respectively, w is the weight vector, b is bias, and g represents the activation function. The feed-forward operations are used after the dropout regularization as follow[11]:where r denotes the vector of independent Bernoulli random variables. Parameter dropout probability (dp), which controls the number of dropped neurons expressed as a percentage, is not trainable and represents CNN’s hyperparameter. Since its value is continuous within the range [0, 1], finding a proper value of this parameter for a specific problem at hand (dataset) is an NP-hard challenge. Finding the proper value of this parameter falls into the category of both CNNs challenges, overfitting avoidance by using dropout regularization and hyperparameter optimization. Metaheuristics, especially nature-inspired ones like swarm intelligence showed as efficient methods for tackling NP-hard problems. Swarm intelligence are population-based, stochastic algorithms that simulate groups of natural organisms such as a flock of birds, fish, and whales, a group of bats and butterflies, colonies of ants and bees, etc. These algorithms perform the search process by investigating within the boundaries of previously discovered parts of the search space (exploitation, intensification) and by exploring novel search regions (exploration, diversification)[29]. Swarm intelligence algorithms have been applied for solving various real-world numerical optimization problems from different domains such as wireless sensor networks (WSNs)[30-33], cloud and edge computing[34,35], image thresholding[36], and many others[37]. The most current and prominent research field is from the domain of hybrid methods between swarm intelligence and machine learning. Researches from this domain have grown rapidly during the past few years and some examples include hyperparameters’ optimization[8,38,39], feature selection[40], predicting time series, e.g. the number of COVID-19 cases[41,42] and ANNs training[43,44].

Proposed method

This section first gives details of the SCA metaheuristics. Afterward, observed shortcomings of its basic version are elaborated. Finally, details of the proposed method that overcomes deficiencies of the basic SCA are provided.

The original SCA method

The inspiration for SCA is taken from the mathematical model of those two important trigonometric functions[12]. Solutions’ positions in the population are updated based on the sine and cosine functions outputs which makes them oscillate around the best solution. These functions return values between − 1 and + 1, which is the mechanism that keeps the solutions fluctuating. An algorithm starts by randomly generating various candidate solutions within the boundaries of the search space in the initialization phase. Exploration and exploitation are controlled differently throughout the execution by random adaptive variables. Solutions’ position update process is performed in each iteration by using the following equations[12]:where and denote the current solution’s position in the i-th dimension at t-th and -th iteration, respectively, are pseudo-randomly generated numbers, the represents destination point’s position (current best approximation of an optimum) in the i-th dimension, while symbol | | denotes the absolute value. These two equations are used in combination by using control parameter :where represents a randomly generated number between 0 and 1. It is noted that for every component of each solution in the population, new values for pseudo-random parameters are generated. Four random parameters control the algorithm’s search process and they influence current and the best solution positions. The balance between solutions is required to efficiently converge towards the global optima and it is achieved by changing the range of the based functions in an ad-hoc manner. The sine and cosine functions exhibit cyclic patterns which allow for reposition around the solution. This behavior guarantees exploitation. The algorithm needs to be enabled to search outside of their corresponding destinations which is possible due to the changes in ranges of sine and cosine functions. Furthermore, the solution requires its position not to overlap with the areas of other solutions. For better quality of randomness, the values for parameter are generated within the range and that guarantees exploration. The Eq. (13) controls the balance between diversification and exploitation.where t is the current iteration, T represents the maximum number of iterations in a run, while a is a constant.

Limitation of basic SCA and proposed hybrid algorithm

The SCA metaheuristics is relatively simple and it does not incorporate many control parameters, yet manages to obtain outstanding results for bound-constrained and constrained benchmarks[12], as well as for various practical challenges[13]. Notwithstanding the good exploitation and exploration performance of original SCA, by executing extensive empirical simulations with standard Congress on Evolutionary Computation (CEC) instances, it was observed that in some runs. the algorithm in later iterations converges to the optimal region and does not have enough cycles to perform there a fine-tuned exploitation. This is mainly because the basic search equation (Eq. 12), either by executed with sine, either by cosine operations, is oriented towards the current best approximation of the optimum () for each solutions’ parameter i. Moreover, notwithstanding that the basic SCA’s search is very efficient in exploitation, there is still some space for improvements. Taking into account observed drawbacks of the original SCA, with an expense of increasing algorithms’ complexity, the following changes were incorporated in the basic SCA: the opposition-based learning (OBL) mechanism is applied to current best solution and search equation from the well-known FA metaheuristics, that proved excellent exploitation capabilities[45,46], is used in the basic SCA search process along with sine and cosine search equations. The OBL mechanism is introduced in[47] and it was proven that it can substantially enhance the metaheuristics search process. For each i-th component of the solution X, the opposite individual is generated by using Eq. 14.where and are lower and upper bounds of component i, respectively. It would be computationally expensive if the OBL is applied to every solution in population in each iteration, so this mechanism is applied only for the current best solution in the following way: in each iteration, the opposite current best solution is created, then if the fitness of is better than the fitness of , the is designated as the new current best . In this way, if in earlier iterations the algorithm misses the optimum domain, there are high chances the current’s best opposite would hit the optimum. On top of the first change, in each iteration the following FA search equation is used with the equal probability along with sine and cosine search mechanisms[48]:where parameter denotes the attractiveness at a distance , marks a randomization parameter, represents a random number drawn from either uniform or Gaussian distribution, and represents the distance between two observed fireflies i and j. It is noted that contrarily to the original FA, the solution is chosen randomly from the population. The proposed method uses dynamic parameter alpha as it was suggested in[48]. In this way, a trade-off between exploitation and exploration is being adjusted in favor of intensification as the iterations progress, as shown in Eq. (16). More details regarding the FA’s control parameters can be captured from[48].With the FA’s search equation, the basic SCA equation 12 is updated as follows:The proposed method is named opposition best SCA firefly search (OBSCA-FS) and its pseudo-code is given in Algorithm 1. Some practical limitations of the proposed method include additional control parameters and more function evaluations in each iteration, which is considered in experiments for a fair comparative analysis.

The OBSCA-FS complexity

The amount of fitness function evaluations (FFEs) is commonly used as a measure to describe the complexity of the swarm intelligence metaheuristics approaches, due to the fact that the evaluation of the objective function represents the most computationally expensive operation, as discussed by[49]. The complexity is typically expressed in terms of FFE over the number of solutions N and number of iterations T. The proposed OBSCA-FS algorithm utilizes just one additional solution evaluation in each iteration during the execution, when the opposite solution of the current best solution is being generated. Consequently, if N denotes the number of solutions, and T represents the number of iterations, the complexity of the OBSCA-FS can be formulated as . Thus, the proposed algorithm slightly increases the complexity of the original algorithm. Moreover, the slight increase in complexity over the basic algorithms is justified because the OBSCA-FS algorithm exhibits significant performance improvements over both basic algorithms, SCA and FA. The enhancements are obvious both for benchmark function set and for the dropout regularization machine learning task, as described in “CEC2019 benchmark simulations” and “CNN dropout regularization simulations” Sections.

Ethics approval

The authors declare that they their work is compliant with ethical standards.

Consent to participate

All authors have given their consent for this research.

Consent for publication

All authors have given their consent for publication of this work.

CEC2019 benchmark simulations

Following proper experimental design, the proposed method was first validated on standard unconstrained benchmarks before it is applied for a real-world problem. In the case of devised OBSCA-FS, a group of 10 relatively novel Congress on Evolutionary Computation 2019 (CEC2019)[50] benchmarks was utilized and simulation results along with comparative analysis and statistical tests are presented in this section. The results of the proposed OBSCA-FS are compared to the original SCA and FA, and other eight state-of-the-art metaheuristics (EHOI, EHO, SSA, GOA, WOA, BBO, MFO, PSO). For the purpose of this research, all above mentioned approaches were implemented and tested. Results of all opponents, including the original SCA, were obtained through extensive simulations. In an effort for this research, original FA was also implemented and tested. Experimental results for the same set of benchmark functions were previously reported in[51]. Nevertheless, simulations presented in this research have been recreated to validate results from[51] and to provide firm grounds for a more objective comparative analysis. Algorithms in[51] were executed by using N=50 and T=500, and this experimental setup could result in biased comparative analysis due to the fact that not all methods use the same amount of FFEs in one iteration. To solve this potential problem, this research utilizes the termination condition that was set according to the total used FFEs for all observed methods. To establish similar experimental preconditions as in[51], the maxFFEs number was set to 25.000 (). The summary of OBSCA-FS control parameters’ adjustments used in experiments is shown in Table 1. The same control parameters were used also in FA’s evaluations.
Table 1

Control parameter summary.

Parameter descriptionValue
Population size N49
Maximum iteration number T500
Parameter \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r_1$$\end{document}r1Changes according to Eq. (13)
Parameter \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r_2$$\end{document}r2\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in [0,2\pi$$\end{document}[0,2π]
Parameter \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r_3$$\end{document}r3\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in [0,1]$$\end{document}[0,1]
Parameter \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r_4$$\end{document}r4\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\in [0,1]$$\end{document}[0,1]
FA’s Absorption coefficient \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\gamma$$\end{document}γ1
FA’s attractiveness parameter at \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r=0$$\end{document}r=0 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta _{0}$$\end{document}β01
FA’s randomization parameter \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha$$\end{document}αChanges according to Eq. (16)
FA’s Initial \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha$$\end{document}α value \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha _{0}$$\end{document}α00.5
Control parameter summary. Obtained experimental results and comparison for mean and standard deviation values for 10 CEC2019 functions are provided in Table 2, where for better readability best results for the mean indicator are marked bold.
Table 2

Result comparison of different well-known metaheuristics on CEC2019 benchmark functions.

FunctionStatsEHOIEHOSCASSAGOAWOABBOMFOPSOFAOBSCA-FS
CEC01Mean4.76E+041.35E+079.83E+093.21E+091.61E+101.03E+103.52E+107.17E+096.75E+117.43E+044.83E+03
Std2.14E+037.91E+066.95E+091.42E+098.99E+99.14E+092.32E+108.69E+092.34E+114.49E+034.21E+03
CEC02Mean1.70E+011.72E+011.75E+011.73E+011.74E+011.73E+018.87E+011.74E+018.56E+022.85E+012.41E+00
Std3.66E−167.29E−155.19E−036.55E−053.23E−021.95E−032.45E+014.17E−153.87E+023.21E+025.32E+01
CEC03Mean1.27E+011.27E+011.27E+011.27E+011.27E+011.27E+011.27E+011.27E+011.27E+011.27E+011.27E+01
Std3.95E−167.44E−163.25E−043.11E−156.47E−047.94E−065.25E−074.38E−054.12E−045.22E−014.03E−01
CEC04Mean1.28E+011.55E+018.32E+023.25E+011.51E+022.65E+026.95E+011.38E+026.92E+013.89E+011.00E+01
Std4.26E+008.52E+003.85E+021.09E+011.13E+021.39E+022.99E+011.15E+025.43E+012.32E−011.19E+00
CEC05Mean1.05E+001.07E+002.23E+001.35E+001.33E+001.67E+001.31E+001.13E+001.55E+001.13E+001.01E+00
Std3.25E−032.41E−027.81E−022.33E−011.21E−013.86E−029.63E−026.56E−021.18E−014.26E−022.17E−02
CEC06Mean8.33E+009.45E+001.04E+013.79E+006.19E+009.14E+005.78E+004.92E+001.03E+011.05E+011.86E+00
Std6.23E−011.31E+008.15E+001.23E+001.33E+001.05E+002.99E−012.13E+003.35E+006.20E−014.46E−02
CEC07Mean1.42E+021.81E+026.38E+022.89E+022.87E+024.53E+024.92E+003.19E+025.97E+024.91E+023.85E+00
Std1.13E+021.51E+022.78E+022.25E+021.75E+022.25E+021.21E+002.15E+021.89E+021.23E+028.36E+01
CEC08Mean2.69E+003.15E+005.77E+005.08E+005.49E+005.75E+004.81E+005.45E+005.10E+002.78E+002.83E+00
Std9.15E−021.44E+007.29E−017.83E−015.14E−017.29E−011.03E+005.62E−017.33E−018.99E−019.13E−01
CEC09Mean2.29E+002.41E+008.75E+012.38E+002.45E+005.16E+003.75E+002.46E+002.65E+004.95E+001.73E+00
Std5.55E−032.18E−025.63E+015.33E−026.41E−025.29E−013.14E−016.76E−028.45E−022.83E−011.54E−02
CEC10Mean1.92E+012.11E+012.08E+012.03E+012.00E+012.05E+012.07E+012.02E+012.06E+012.02E+011.32E+01
Std3.49E+007.29E+006.45E+008.19E+006.67E+003.52E−017.13E−006.66E−019.81E+029.13E−021.56E−02
Result comparison of different well-known metaheuristics on CEC2019 benchmark functions. Reported results in Table 2 communicate the superior performance of proposed OBSCA-FS when compared to other state-of-the-art approaches. For almost all instances, the OBSCA-FS managed to establish the best mean value. The only exceptions are CEC03 benchmark, where all approaches obtained the same mean indicator value and CEC08 in which case the results reported for EHOI are the best, while original FA performed slightly better than OBSCA-FS. Furthermore, based on experimental data it can be concluded that the OBSCA-FS hybrid method substantially improves the performance of both algorithms, SCA and FA, which is at the same time the basic goal of devising hybrid methods. However, when comparing different methods it is not enough just to state that one method is better than the other in terms of results, it also should be determined whether the enhancements are statistically significant. Following this assumption, a Friedman test[52,53] and two-way analysis of variance by ranks is performed to validate the significant difference between proposed OBSCA-FS and other adversary metaheuristics. The Friedman test results for eleven approaches applied to 10 functions are provided in Table 3.
Table 3

Friedman ranks for the comparable method over 10 CEC2019 instances.

FunctionEHOIEHOSCASSAGOAWOABBOMFOPSOFAOBSCA-FS
CEC012475981061131
CEC024484741041191
CEC0366666666666
CEC042311491078651
CEC052.52.511681075941
CEC066811257431091
CEC073411568271091
CEC081411781059623
CEC092311459867101
CEC10291134.574.571071
Average3.054.759.84.66.757.96.356.18.66.41.7
Rank2411389651071
Friedman ranks for the comparable method over 10 CEC2019 instances. From Table 3 can be noticed that the proposed OBSCA-FS is advanced in terms of performance than the other 10 algorithms with an average rank of 1.7. The basic FA and SCA have average rankings of 6.4 and 9.8, respectively. Also, the Friedman statistics () is greater than the critical value with 10 degrees of freedom (18.3), at significance level , therefore the null hypothesis () is rejected and it can be concluded that the proposed OBSCA-FS obtains results which are significantly different than other 10 methods. However, in[54] is reported that the Iman and Davenport’s test[55] may be more precise than the having this in mind, Iman and Davenport’s test was also executed. Calculated statistic of , which is greater than the F-distribution critical value (). Therefore, Iman and Davenport’s test also rejects . In the case of both tests, the is less than the significance level. Summary of results for both statistical tests is given in Table 4.
Table 4

Friedman and Iman-Davenport test results ().

Friedman value\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\chi ^2$$\end{document}χ2 critical valuep-valueIman-Davenport valueF critical valuep-value
5.12E+011.83E+011.11E−169.46E+001.93E+001.11E−13
Friedman and Iman-Davenport test results (). As the null hypothesis is rejected by Friedman and Iman and Davenport tests, we proceeded with the non-parametric post-hoc procedure, with Holm’s step-down procedure, and the obtained results are reported in Table 5.
Table 5

Holm’s step-down procedure result.

Comparisonp-valueRank0.05/(k-i)0.1/(k-i)
OBSCA-FS versus SCA2.37E−0800.0050000.01000
OBSCA-FS versus PSO1.46E−0610.0055560.01111
OBSCA-FS versus WOA1.46E−0520.0062500.01250
OBSCA-FS versus GOA3.31E−0430.0071430.01429
OBSCA-FS versus FA7.66E−0440.0083330.01667
OBSCA-FS versus BBO8.59E−0450.0100000.02000
OBSCA-FS versus MFO1.50E−0360.0125000.02500
OBSCA-FS versus EHO1.98E−0270.0166670.03333
OBSCA-FS versus SSA2.52E−0280.0250000.05000
OBSCA-FS versus EHOI1.81E−0190.0500000.10000
Holm’s step-down procedure result. Table 5 shows that the proposed method significantly outperformed all compared methods at significance level , as well as all algorithms except EHOI at significance level . Figure 1 illustrates a head-to-head convergence comparison between the proposed OBSCA-FS, second best approach (EHOI) and relevant basic metaheuristics, with respect to the results on the 10 benchmark functions. Since OBSCA-FS is a hybrid of SCA and FA, convergence graphs for those two methods were included as well. The plots illustrate how the fitness evaluation decreases over the course of iterations for each test function in turn.
Figure 1

Convergence speed graphs of the 10 CEC 2019 benchmark functions as direct comparison between proposed OBSCA-FS and relevant metaheuristics.

It is important to notice from the convergence graphs that the proposed algorithm combines the best elements of the FA and SCA metaheuristics. It is obvious from the Fig. 1 that OBSCA-FS converges faster from both FA and SCA, therefore establishing itself as more efficient metaheuristics. However, according to the no free lunch theorem, there is no general metaheuristics that will be a perfect solution for all problems. This can be seen on convergence graph for the benchmark F8, where basic FA obtained better results than the proposed OBSCA-FS. Nevertheless, this behavior is normal and expected, as there is always a trade-off. Convergence speed graphs of the 10 CEC 2019 benchmark functions as direct comparison between proposed OBSCA-FS and relevant metaheuristics.

CNN dropout regularization simulations

This section is divided into two parts. In the first part, research findings and comparative analysis from experiments with four standard benchmark datasets are shown, while the second part provides details for dropout regularization experiments with specific MRI dataset.

Dropout simulations for benchmark datasets

The model used in conducted experiments is the same as the one proposed in the referenced paper[7], where the performance of distinguished bat algorithm (BA), cuckoo search (CS), FA, and particle swarm optimization (PSO) swarm intelligence metaheuristics for dropout regularization challenge was reported. In this manuscript, the same experimental environment as in[7] was established due to the fair comparative analysis. As acclaimed in “Introduction” Section, the potential of swarm intelligence approaches for this problem was not fully investigated and the aim of this experiment is to further examine its potential in this domain, as well as to validate the proposed state-of-the-art hybrid OBSCA-FS method against this practical challenge. The testing framework was developed in Python with its core and data science libraries and API’s keras, scikitlearn, numpy, scipy along with pandas and matplotlib for visualization. The machine that was used for testing has a 6 NVIDIA GTX 1080 GPU with Intel CoreTM i7-8700K CPU, 64GB RAM, and Windows 10 OS. For validation, 4 standard datasets were used: MNIST (http://yann.lecun.com/exdb/mnist/), Semeion (https://archive.ics.uci.edu/ml/datasets/Semeion+Handwritten+Digit), USPS (http://statweb.stanford.edu/tibs/ElemStatLearn/datasets/zip.info.txt) and CIFAR10 (http://www.cs.toronto.edu/kriz/cifar.html). For more details regarding characteristics and the number of available observations in each dataset, please refer to the provided links. Two different CNN architectures as in[7], as it is provided in the default Caffe examples, are utilized: one for MNIST, Semeion, and USPS and one for the CIFAR-10 dataset. The only difference is that in both models additional dropout layer is added and in the case of USPS and Semeion datasets, the kernel size of instead of is used for convolution layers because these two datasets have lower resolution. Example instances of two models are given in Figs. 2 and 3.
Figure 2

Example instance of MNIST, Semeion and USPS model.

Figure 3

Example instance of CIFAR10 model.

Example instance of MNIST, Semeion and USPS model. Example instance of CIFAR10 model. In all simulations besides the dropout probability (dp) for the dropout layer, which is subject to optimization, L1 regularization (penalty) and L2 regularization (weight decay) were also employed. For training the models, RMSProp optimizer was executed with a learning rate . The parameters’ tuple (, , ) was fixed and only dp was optimized. Therefore, solutions’ encoding is very straightforward and every solution consists of only one parameter with values . The classification error rate is used as the fitness function, therefore a problem is formulated as a minimization challenge. With the goal of visualizing proposed methodology, general OBSCA-FS and fitness calculation flow-charts are shown in Fig. 4.
Figure 4

Proposed methodology for dp regularization.

Proposed methodology for dp regularization. Additionally, with the goals of straightforward comparative analysis, in the context that all metaheuristics are tested under the same conditions, and of establishing referenced (baseline) models, standard Caffe architecture with and without dropout is also included in comparison and for other metaheuristics, standard default Caffe parameters for , , and were used. These parameters over utilized datasets are summarized in Table 6.
Table 6

The CNN , and adjustments for simulations.

Dataset\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\eta$$\end{document}η\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha$$\end{document}α\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\lambda$$\end{document}λdp
CIFAR-100.0010.90.004[0, 1]
MNIST0.010.90.00005[0, 1]
Semeion0.0010.90.00005[0, 1]
USPS0.010.90.00005[0, 1]
The CNN , and adjustments for simulations. All datasets were split into the training, validation, and testing sets. For determining the fitness of each individual, the classification accuracy of the test set is used. In all experiments, the loss function is employed and for the CIFAR-10 dataset model is trained in 4,000, while for the remaining 3 datasets, 10,000 epochs are used for training. Model details and the number of samples along with the batch size are summarized in Table 7.
Table 7

Configuration of the datasets used in experiments.

DatasetTraining set samples (batch size)Validation set samples (batch size)Testing set (batch size)Epochs
CIFAR-1020,000 (100)30,000 (100)10,000 (100)4000
MNIST20,000 (64)40,000 (100)10,000 (100)10,000
Semeion200 (2)400 (400)993 (993)10,000
USPS2406 (32)4885 (977)2007 (2007)10,000
Configuration of the datasets used in experiments. For the purpose of this experiment, all approaches that were included in the comparative analysis were implemented and tested. Control parameters for BA, FA, CS, and PSO that were used in the comparative analysis can be retrieved from[7]. Proposed OBSCA-FS was tested with the same parameters as presented in Table 1. Moreover, as already stated, to further investigate swarm algorithms’ performance for tackling dropout regularization issue, the EHO, WOA, SSA, GOA, BBO and SCA metaheuristics were also implemented and tested with the suggested parameters for bound-constrained optimization in relevant publications. The control parameters values are summarized in Table 8. All metaheuristics approaches were executed with a total number of 77 FFEs. The study proposed in[7] evaluated methods with and , which also yielded a total of 77 FFEs (7 + 7 10). Cross-validation with 20 runs was carried out for the purpose of providing statistical Friedman’s non-parametric test and average results over 20 independent runs are reported.
Table 8

Configuration of control parameters of the metaheuristics that were implemented and included in the comparative analysis.

AlgorithmControl parameters and their values
BA[56]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$f_{min} = 0$$\end{document}fmin=0, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$f_{max} = 2$$\end{document}fmax=2, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$A = 0.5$$\end{document}A=0.5, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$rand = 0.5$$\end{document}rand=0.5
CS[57]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta = 1.5$$\end{document}β=1.5, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$p = 0.25$$\end{document}p=0.25, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha = 0.8$$\end{document}α=0.8
PSO[58]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$c_1 = 1.7$$\end{document}c1=1.7, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$c_2 = 1.7$$\end{document}c2=1.7, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$w = 0.7$$\end{document}w=0.7
EHO[59]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$no_clan = 5$$\end{document}noclan=5, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha = 0.5$$\end{document}α=0.5, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta =0.1$$\end{document}β=0.1, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$no_elite = 2$$\end{document}noelite=2
WOA[60]Initial value of \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$a = 2.0$$\end{document}a=2.0, linearly decreasing to 0
SSA[61]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$c_{1}$$\end{document}c1 non-linearly decreasing from 2 to 0, c2 and c3 rand from [0,1]
GOA[62]c linearly decreasing from 1 to 0
BBO[63]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$hmp= 1$$\end{document}hmp=1, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$imp= 0.1$$\end{document}imp=0.1, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$nbhk = 2$$\end{document}nbhk=2
FA[48]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\gamma = 1.0$$\end{document}γ=1.0, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta _0 = 1.0$$\end{document}β0=1.0, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha = 0.2$$\end{document}α=0.2
SCA[12]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$a=2$$\end{document}a=2, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r_{1}$$\end{document}r1 linearly decreasing from 2 to 0
Configuration of control parameters of the metaheuristics that were implemented and included in the comparative analysis. The average value of accuracy along with the mean value of dp obtained in the MNIST, Semeion, USPS and CIFAR-10 datasets simulations are reported in Table 9. Best obtained accuracy among metaheuristics-based approaches is marked bold. From the presented table, it can be seen that the average accuracy is not consistent over the observed datasets. This is mostly because of the different and distinctive nature of each dataset included in the experiments, in terms of the overall number of images in the dataset, the amount of features, and the contents of images.
Table 9

Comparative results of the suggested OBSCA-FS method and other metaheuristics approaches in terms of mean classification accuracy.

MethodMNISTSemeionUSPSCIFAR-10
acc.dpacc.dpacc.dpacc.dp
Caffe99.07097.62095.80071.470
Dropout Caffe99.180.598.140.596.210.572.080.5
BA99.140.49198.350.69296.450.76271.490.633
CS99.140.48998.210.54496.310.71571.210.669
PSO99.160.49397.790.37196.330.72571.510.621
EHO99.130.47598.110.48196.240.68271.150.705
WOA99.150.48998.230.56196.320.72271.230.685
SSA99.190.49998.310.64296.410.75371.580.529
GOA99.160.49298.150.51396.150.48170.950.849
BBO99.130.47498.160.51596.170.48371.080.768
FA99.180.49598.290.61996.420.75871.550.583
SCA99.170.49698.250.58096.290.70571.540.597
OBSCA-FS99.280.52498.480.72296.930.83872.540.394
Comparative results of the suggested OBSCA-FS method and other metaheuristics approaches in terms of mean classification accuracy. The findings presented in Table 9 point out the superior performance of the suggested OBSCA-FS approach in regard to the dp value that has been subjected to the optimization. On the MNIST dataset, the suggested OBSCA-FS approach achieved the superior accuracy of 99.28% with obtained dp value of 0.524. On the same dataset, other metaheuristics algorithms determined the dp value below the value of the standard Dropout Caffe . In this scenario, the findings indicate that the dp value should be just over 0.5 to obtain better accuracy values, and the suggested OBSCA-FS approach was the only algorithm that has achieved it. On the Semeion dataset, the suggested OBSCA-FS algorithm achieved the best accuracy of 98.48%, by obtaining the dp value of 0.722. In this case, it can be concluded that the accuracy increases with values of dp, that are over the standard Dropout Caffe value 0.5. The runner-up approach was BA, that was able to achieve the accuracy of 98.35% by . The basic Caffe method that does not utilize the dropout () was able to achieve 97.62% accuracy, while the Dropout Caffe () obtained the accuracy of 98.14%. Similar results are observed on the USPS dataset as well. The suggested OBSCA-FA method again obtained the best accuracy of 96.93% with determined dp value of 0.838. Similarly to the first two observed datasets, the accuracy increases with the increase of dp value. The second best result has been obtained by BA, that achieved 96.45 % by . The basic Caffe and Dropout Caffe were significantly behind the proposed OBSCA-FA, with the accuracy lesser by approx. 1.1% and 0.7%, respectively. Lastly, in the case of CIFAR-10 dataset, a different pattern can be observed. The findings indicate that, if the dp is greater than the Dropout Caffe (), the performance starts to decrease, while the accuracy drops down. In other words, the network drops out neurons, and it starts to lose the ability to generalize well. On the other side, if the dp value is too small, again, the accuracy will decrease (similar to the basic Caffe with ). The best performances for the CIFAR-10 dataset are achieved for the dp values slightly lesser than 0.5. The proposed OBSCA-FA approach scored the best accuracy of 72.54% with the . Moreover, it was the only approach that was able to find the dp value below 0.5 boundary, while other metaheuristics approaches got stuck with the dp values in range [0.5, 1]. Finally, it is worth noticing that the OBSCA-FS approach, as a hybrid between SCA and FA, significantly outperformed both basic metaheuristics versions in all performed tests. This way, the enhancements of the OBSCA-FS over the original implementations of SCA and FA, observed on the unconstrained benchmark functions tests, were also confirmed on the practical task of optimizing the dropout regularization. Average convergence speed graphs for implemented OBSCA-FS, SCA, FA and three other representative metaheuristics (PSO, BA and SSA) for mean classification error for MNIST, CIFAR-10, Semeion, and USPS datasets, generated over 20 independent runs with 77 FFEs, are provided in Fig. 5.
Figure 5

Convergence graph of MNIST, CIFAR-10, Semeion and USPS datasets for average classification error for OBSCA-FS, SCA and WOA.

Convergence graph of MNIST, CIFAR-10, Semeion and USPS datasets for average classification error for OBSCA-FS, SCA and WOA. Similarly, as in unconstrained experiments, a Friedman test and two-way analysis of variance by ranks are performed. Results are shown in Table 10. From presented statistical tests can be observed that the proposed OBSCA-FS established with the rank 1, the SSA proved as the second-best metaheuristics, while the rank of 3 obtained the original FA metaheuristics.
Table 10

Friedman ranks for the comparable method over 4 CNN classification instances.

FunctionCaffeDropout CaffeBACSPSOEHOWOASSAGOABBOFASCAOBSCA-FS
MNIST133.59.59.56.511.5826.511.53.551
Semeion13102712116398451
USPS13102759641211381
CIFAR-1082710611931312451
Average11.756.3755.1258.3757.37510.6257.25310.12510.6253.6255.751
Rank13649811721012351
After completing the necessary calculations, the Iman and Davenport test score was 36.95, and it was compared to the F-distribution critical value (). As the test returns considerably larger value, this test rejects . Additionally, the Friedman statistics () is greater than the critical value with ten degrees of freedom (1, 82), with the significance level . Finally, the null hypothesis () can be rejected. This indicates that OBSCA-FS performances were significantly superior over the rest of the metaheuristics included in the experiments. Friedman ranks for the comparable method over 4 CNN classification instances. Lastly, although not being the main topic of this research described in this paper, another mini experiment has been conducted, to further evaluate the performances of the proposed OBSCA-FS algorithm. In this additional experiment, all four parameters from the Table 6, namely dp (dropout probability), (learning rate), (L1 regularization) and (L2 regularization - weight decay), have been subjected to the optimization process, without taking the standard default Caffe parameter values. The parameter ranges that were used in the experiments were , , and . The experiment utilized the same setup, the same number of runs and the same amount of training as the main benchmark experiments described above. The accuracy results obtained by the basic SCA were as follows: 99.19% on the MNIST dataset, 98.28% on the Semeion dataset, 96.33% on the USPS dataset, and 71.56% on the CIFAR-10 dataset. On the other hand, the accuracy values achieved by the proposed OBSCA-FS method were as follows: 99.32% in the case of MNIST dataset, 98.55% on the Semeion dataset, 97.03% on the USPS dataset, and finally, 72.69% on the CIFAR-19 dataset. The improvement of the accuracy results for both the basic SCA and the proposed OBSCA-FS methods are minor, leading to the conclusion that the default values for , and used by the Caffe are well adjusted. However, in case of all four parameters being subjected to the optimization process, every metaheuristics solution is being encoded with four continuous parameters. This significantly increases the complexity of the method, while the minor improvements of the accuracy are not justifying it.

Dropout simulations for MRI benchmark dataset

In a similar way as presented in[64], MRI dataset was used to further validate the proposed approach. The benchmark dataset (https://figshare.com/articles/dataset/brain_tumor_dataset/1512427) is comprised of 3064 T1-weighted MRIs with Glioma, Meningioma, and Pituitary brain tumor classes, obtained from 233 subjects by Cheng et. al[65]. This research employed the same image processing as in[64]: pixel values for all images were normalized to scale [0, 1], followed by the data augmentation approach to increase the volume of the training set. The dimensions of images were set to pixels. Data augmentation step included the generation of fake data by randomly modifying the original images, and the addition of this generated data to the initial dataset. In this research, random modifications included rotating the original image by 10, 20, or 30 degrees in a random direction, translating the image by 15 pixels, resizing the original image, mirroring, and finally, using combinations of modifications at once. The dataset was originally comprised of 3064 axial images divided into three classes: 708 meningioma, 1426 glioma, and 930 pituitary tumor images. Upon completion of the data augmentation process, every category consisted of 1521 images that were utilized for training phase, and 115 that were used for the testing phase, with the total amount of images being 4908. Additional information about the image pre-processing and dataset splitting can be obtained from[64]. The same fitness function formulation as in simulations for standard datasets is used and for more details refer to “Dropout simulations for benchmark datasets” Section. The proposed OBSCA-FS algorithm was tested for the dropout rate and validated against several cutting-edge metaheuristics approaches. The utilized CNN topology was derived from previous research published in[66], that was dealing with the hyperparameters’ optimization for this particular dataset. The CNN structure shown in Fig. 6 obtained the best results in[66], and it was consequently utilized in this research for testing and optimizing the dropout probability parameter. Finally, the CNN uses the Adam optimizer and the learning rate , again as the product of research[66].
Figure 6

The CNN structure utilized for MRI dataset.

The CNN structure utilized for MRI dataset. Research published in[64] is evolving 50 () candidate CNN structures in 15 rounds , that for most metaheuristics algorithms corresponds to a total number of 800 FFEs (). However, due to the fact that not all metaheuristics algorithms utilize the same amount of FFEs in every round of execution, in the proposed research was used as the terminating condition. The control parameters for opposing metaheuristics approaches were obtained from the original publications and are shown in Table 11.
Table 11

Control parameters’ values for metaheuristics methods included in the experiments.

MetaheuristicsParameters’ values
GA[64]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$p_{c} = 0.5$$\end{document}pc=0.5, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$p_{m} = 0.2$$\end{document}pm=0.2
FA[48]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha = 0.5$$\end{document}α=0.5, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta = 0.2$$\end{document}β=0.2, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\gamma = 1.0$$\end{document}γ=1.0
mFA[39]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha = 0.5$$\end{document}α=0.5, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta = 0.2$$\end{document}β=0.2, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\gamma = 1.0$$\end{document}γ=1.0, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$TL = 20 \, FFEs$$\end{document}TL=20FFEs
BA[56]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Q_{min} = 0$$\end{document}Qmin=0, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$Q_{max} = 2$$\end{document}Qmax=2, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$A=0.5$$\end{document}A=0.5, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r=0.5$$\end{document}r=0.5
EHO[59]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$no_clan = 5$$\end{document}noclan=5, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\alpha = 0.5$$\end{document}α=0.5, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta =0.1$$\end{document}β=0.1, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$no_elite = 2$$\end{document}noelite=2
WOA[60]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$a_{1}$$\end{document}a1 linearly decreasing from 2 to 0, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$a_{2}$$\end{document}a2 linearly decreasing from -1 to -2, b=1
SCA[12]\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$a=2$$\end{document}a=2, \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$r_{1}$$\end{document}r1 linearly decreasing from 2 to 0
The proposed OBSCA-FS driven CNN was also compared to other cutting-edge, non-metaheuristics approaches, including SVM + RFE, Vanilla preprocessing + shallow CNN, LeNet-5, VGG19 and DenseNet. Since those models are considered to be state of the art, they were tested with default parameters. Control parameters’ values for metaheuristics methods included in the experiments. MRI tumor grades classification comparative analysis. The simulation results are presented in the Table 12, where the best accuracy and dropout probability have been reported. The presented findings indicate that the proposed OBSCA-FS approach obtained the best average accuracy and the dropout probability of 0.43, clearly outperforming all traditional and metaheuristics approaches included in the research.
Table 12

MRI tumor grades classification comparative analysis.

ApproachAccuracy (%)Dropout
SVM + RFE[67]71.2
Vanilla preprocessing + shallow CNN[68]91.4
CNN LeNet-5[69]74.9
VGG19[70]92.6
DenseNet[71]92.7
CNN + GA[64]94.90.33
CNN + mFA[39]96.90.39
CNN + BA[56]95.60.37
CNN + EHO[59]94.80.31
CNN + WOA[60]95.50.36
CNN + HHO[66]96.50.38
CNN + eHHO[66]98.30.41
CNN + FA[39]96.10.37
CNN + SCA[12]96.80.40
CNN + OBSCA-FS98.60.43
Figure 7 depicts the box plots for all approaches that were taken in the comparative analysis. It can be noted that the proposed OBSCA-FS approach obtained the best solution diversity. In other words, it has the highest stability over 10 runs, with the smallest standard deviation. Additionally, once more the proposed OBSCA-FS significantly outperformed both SCA and FA methods. Figure 8 shows normalized confusion matrix for two best approaches on the MRI dataset, namely the CNN + OBSCA-FS and CNN + eHHO.
Figure 7

MRI dataset—best solutions diversity over 10 runs results.

Figure 8

Confusion matrix for OBSCA-FS and eHHO, the two approaches that scored the best results on the MRI dataset.

MRI dataset—best solutions diversity over 10 runs results. Confusion matrix for OBSCA-FS and eHHO, the two approaches that scored the best results on the MRI dataset.

Conclusion

The research presented in this manuscript proposes a novel automated approach for selecting the regularization dropout parameter dp in CNN’s by utilizing the hybridized SCA metaheuristics. The basic SCA is hybridized with the FA’s search equation, moreover, the opposite best solution is generated in each iteration to improve the algorithm’s exploration abilities. The proposed OBSCA-FS method was first evaluated on 10 recent CEC2019 bound-constrained benchmarks and compared with other state-of-the-art approaches tested with the same experimental data. Reported results, as well as conducted statistical tests, deliver the proof that the proposed method performs significantly better than other approaches. Moreover, it was shown that the OBSCA-FS outscores basic SCA and FA metaheuristics. Performance of OBSCA-FS was further validated on practical CNN’s application for optimizing dropout probability value, which is very important in preventing overfitting, as one of the most distinguished challenges from the machine and deep learning area. Reported classification accuracy over MNIST, CIFAR-10, Semieon, and USPS datasets clearly shows that the proposed OBSCA-FS has great potential in this domain. The second experiment included the OBSCA-FS practical implementation for MRI classification. The obtained results confirmed the performances of the proposed method as superior. Due to the great potential of the introduced OBSCA-FS algorithm, in future research, it will be tested on other machine learning challenges and adapted for solving other practical NP-hard problems from the real-world environment. Moreover, the CNN’s regularization will be tackled further by using OBSCA-FS and other similar approaches by fine-tuning other parameters like , and .
  9 in total

1.  Subject independent facial expression recognition with robust face detection using a convolutional neural network.

Authors:  Masakazu Matsugu; Katsuhiko Mori; Yusuke Mitari; Yuji Kaneda
Journal:  Neural Netw       Date:  2003 Jun-Jul

2.  Face recognition: a convolutional neural-network approach.

Authors:  S Lawrence; C L Giles; A C Tsoi; A D Back
Journal:  IEEE Trans Neural Netw       Date:  1997

3.  Classification of brain tumor type and grade using MRI texture and shape in a machine learning scheme.

Authors:  Evangelia I Zacharaki; Sumei Wang; Sanjeev Chawla; Dong Soo Yoo; Ronald Wolf; Elias R Melhem; Christos Davatzikos
Journal:  Magn Reson Med       Date:  2009-12       Impact factor: 4.668

4.  Enhanced Performance of Brain Tumor Classification via Tumor Region Augmentation and Partition.

Authors:  Jun Cheng; Wei Huang; Shuangliang Cao; Ru Yang; Wei Yang; Zhaoqiang Yun; Zhijian Wang; Qianjin Feng
Journal:  PLoS One       Date:  2015-10-08       Impact factor: 3.240

5.  Predicting clustered weather patterns: A test case for applications of convolutional neural networks to spatio-temporal climate data.

Authors:  Ashesh Chattopadhyay; Pedram Hassanzadeh; Saba Pasha
Journal:  Sci Rep       Date:  2020-01-28       Impact factor: 4.379

6.  COVID-19 cases prediction by using hybrid machine learning and beetle antennae search approach.

Authors:  Miodrag Zivkovic; Nebojsa Bacanin; K Venkatachalam; Anand Nayyar; Aleksandar Djordjevic; Ivana Strumberger; Fadi Al-Turjman
Journal:  Sustain Cities Soc       Date:  2020-12-30       Impact factor: 7.587

7.  A comprehensive survey of sine cosine algorithm: variants and applications.

Authors:  Asma Benmessaoud Gabis; Yassine Meraihi; Seyedali Mirjalili; Amar Ramdane-Cherif
Journal:  Artif Intell Rev       Date:  2021-06-02       Impact factor: 8.139

8.  Firefly algorithm for cardinality constrained mean-variance portfolio optimization problem with entropy diversity constraint.

Authors:  Nebojsa Bacanin; Milan Tuba
Journal:  ScientificWorldJournal       Date:  2014-05-29
  9 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.