Literature DB >> 35789599

Non-smooth Bayesian learning for artificial neural networks.

Mohamed Fakhfakh1,2, Lotfi Chaari2, Bassem Bouaziz1, Faiez Gargouri1.   

Abstract

Artificial neural networks (ANNs) are being widely used in supervised machine learning to analyze signals or images for many applications. Using an annotated learning database, one of the main challenges is to optimize the network weights. A lot of work on solving optimization problems or improving optimization methods in machine learning has been proposed successively such as gradient-based method, Newton-type method, meta-heuristic method. For the sake of efficiency, regularization is generally used. When non-smooth regularizers are used especially to promote sparse networks, such as the ℓ 1 norm, this optimization becomes challenging due to non-differentiability issues of the target criterion. In this paper, we propose an MCMC-based optimization scheme formulated in a Bayesian framework. The proposed scheme solves the above-mentioned sparse optimization problem using an efficient sampling scheme and Hamiltonian dynamics. The designed optimizer is conducted on four (4) datasets, and the results are verified by a comparative study with two CNNs. Promising results show the usefulness of the proposed method to allow ANNs, even with low complexity levels, reaching high accuracy rates of up to 94 % . The proposed method is also faster and more robust concerning overfitting issues. More importantly, the training step of the proposed method is much faster than all competing algorithms.
© The Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2022.

Entities:  

Keywords:  Artificial neural networks; Hamiltonian dynamics; Machine learning; Optimization

Year:  2022        PMID: 35789599      PMCID: PMC9244188          DOI: 10.1007/s12652-022-04073-8

Source DB:  PubMed          Journal:  J Ambient Intell Humaniz Comput


Introduction

Machine learning (ML) (Shakshuki et al. 2020) is a subfield of artificial intelligence (AI). It has grown at a remarkable rate, attracting a great number of researchers which are intended to study how a system can perform a task through learning. In fact, an ML system does not follow instructions, but learns from experience, for example, make predictions or decisions by learning from data and keep improving the performance by examining more data. ML research achieved outstanding results on several complex cognitive tasks, including Computer Vision (Alsarhan et al. 2021), Medical diagnoses (Chaabene et al. 2021; Sree et al. 2021), Signal Processing (Jaini et al. 2021), etc. During the last two decades, Deep Learning (DL) architectures (Devunooru et al. 2021; Goyal and Singh 2021) have demonstrated their ability to deal with more voluminous and complex data. Moreover, it has gradually become the most widely used computational approach in the field of ML, achieving outstanding results on several cognitive tasks, matching or even beating those reached by human performance. One of the benefits and difficulties at the same time of DL is the ability to learn from massive amounts of data. In the same sense, Convolutional neural networks (CNN) (Drewek-Ossowicka et al. 2021; Sajja and Kalluri 2021; Fakhfakh et al. 2020a; Ostad-Ali-Askari et al. 2017; Ostad-Ali-Askari and Shayan 2021) are one of the state-of-art deep learning techniques. CNNs are designed to automatically and adaptively learn spatial hierarchies of features through backpropagation (Rumelhart et al. 1986) by using multiple building blocks, such as convolution layers, pooling layers, and fully connected layers. However, training a CNN is a challenging task, especially for deep architecture involving a high number of parameters (model weights) to be estimated. Sophisticated optimization algorithms need therefore to be used. This is indeed the key step in order to fit a given architecture for learning data to minimize the error between ground truth and estimates. In this sense, many optimization algorithms have been proposed in recent years. The performance of the algorithms strongly depends on the convexity and differentiability of the target loss function. Hence, choosing an optimization strategy that seeks to find the global optima in the learning stage is generally challenging. A non-appropriate optimization technique may for instance lead the network to lie at a local minimum during the training phase. Speeding up the optimization process is also a challenging issue for large databases. All learning models, particularly Deep Neural Networks (DNN), are well known for being overparameterized, in fact, relatively few network weights are actually necessary to accurately learn data characteristics. They are also known to have many redundant parameters (Scardapane et al. 2017; Cheng et al. 2015). Moreover, the large number of weight parameters often leads to a heavy memory cost and computation resources. To address this computational issue, several strategies for reducing the number of network weights (weight sparsification) have been proposed, either on pre-trained models or during the training phase. To promote sparsity of DNNs, three main categories of methods can be identified: pruning, dropout, and sparse optimization based techniques. Although optimization-based sparsification is the most promising class, introducing sparse regularizers generally leads to non-differentiable cost functions. Using gradient-based techniques is therefore sub-optimal. Moreover, non-convex regularizers such as the one, are more likely to produce unbiased model with sparser solutions. However, to the best of our knowledge, there is no previous work proposing a flexible optimization-based technique allowing to handle convex and non-convex regularizers, with non-differentiable cost functions. The originality of the present work lies in :In this context, the use of Bayesian techniques has made huge strides in a variety of disciplines over the decades, and there are many practical advantages. The core concept is to use a probabilistic formulation to integrate all uncertainties throughout the model. Resorting to Bayesian techniques in our case is motivated by the ability of these methods to incorporate flexible prior models translating these uncertainties, while reducing the user input as all the model parameters and hyperparameters can be estimated from the data. Moreover, Bayesian inference using Markov Chain Monte Carlo (MCMC) methods (Chaari et al. 2014, 2016) guarantees insensitivity to local minima issues. The target space can be fully explored once sufficient mixing properties are enjoyed by the sampled chains. Such methods can also be used as an alternative to variational methods for non-convex problems, where standard optimization techniques still suffer from computational and convergence limitations. Pruning removes weight parameters that are insensitive to the performance of established dense networks. The main drawback is linked to the pruning criteria which requires manual setups of layer sensitivity. Heuristic assumptions are also necessary throughout the pruning process (Han et al. 2015; Anwar et al. 2017). Dropout reduces the size of networks during training by randomly dropping units along with their connections from DNNs. This method can reduce overfitting efficiently and improve the performance. Nonetheless, training a Dropout network, usually takes more time than training a standard neural network (Srivastava et al. 2014). Optimization-based methods promote sparsity in networks by introducing a regularization term into the cost function (loss) that promotes estimation with a large number of zero weights. Sparse neural networks are being widely investigated for applications (Fan et al. 2020; Han et al. 2017) and could even achieve better performance than their original networks. The Bayesian formulation of the optimization problem for DNNs. It is a powerful strategy for finding the extrema of objective functions that are expensive to evaluate. The proposed flexible and efficient optimization procedure which solves the non-differentiability problem and allows handling different regularizers. The proposed method ensures convergence to the global minimum of the formulated cost-function in contrast to other techniques (for example gradient based). As mentioned above, the goal of this paper is to develop a Bayesian model to minimize the target non-linear cost function. The main contribution lies in the adaptation of non-smooth Hamiltonian methods to fit sparse ANNs as weights are subject to sparsity constraints. Under the Bayesian formulation, this can be modeled using a Laplace prior distribution. On the other hand, despite the above-mentioned advantages of Bayesian formulations and MCMC-based inference schemes, these techniques remain time-consuming. Moreover, non-smooth priors such as the Laplace one make complicated the use of standard sampling methods. This also holds for gradient-based techniques when regularizations are used. It is worth noting that regularization is typically used for signal and image recovery problems where the target data enjoys high sparsity levels, either in the original or a transform space (Loris et al. 2007). The problem is also used in the artificial neural networks literature under both constrained (Gen et al. 2020) and Lagrangian (Ashwini and Shital 2019) formulations. In this sense, non-smooth Hamiltonian methods allow designing of fast and efficient sampling schemes while handling non-differentiable energy functions. Our contribution is therefore focused on a new Bayesian optimization scheme adapted to train sparse neural networks without user configuration. The rest of this paper is organized as follows. After an introduction covering the context of this research, Sect. 2 is dedicated to the state of the art. Then, the addressed problem is formulated in Sect. 3. The proposed efficient Bayesian optimization scheme is developed in Sect. 4 and validated in Sect. 5. Therefore, The discussion will be presented in Sect. 6. Finally, The conclusion and future work are drawn in Sect. 7.

Related work

The performance of a DL algorithm mainly depends on the optimization procedure used during the learning process. The essence of most architectures is to build an optimization model and learn the parameters from the available learning data. Although there is not a single solution to find the optimal set of parameters (i.e. weights) for a neural network with reasonable complexity, several researches have focused on the improvement of optimization algorithms in order to enhance the efficiency of deep learning architectures in terms of accuracy, robustness and convergence time. Indeed, optimization methods can be divided into three categories (Sun et al. 2019; Zaheer and Shaziya 2019): (i) first-order optimization methods such as stochastic gradient; (ii) high-order optimization methods, mainly Newton’s algorithm; and (iii) heuristic/meta-heuristic derivative-free optimization methods. First-order optimization algorithms (Xie and Zhang 2021), minimize an objective function parameterized by a model’s weight by updating the weights in the opposite direction of gradients of the objective function. When non-convex or non-differentiable functions are used, these methods may suffer from slow convergence and local minima issues. The Stochastic Gradient Descent (SGD) method (Robbins and Monro 1951; Sutskever et al. 2013) is one of the core techniques behind the success of deep neural networks since it alleviates the above-mentioned difficulties. Therefore, the limitation is using equal-sized steps for all parameters, irrespective of gradient behavior. Adaptive moment estimation (Adam) (Kingma and Ba 2014) is one of the recent and popular variants. It allows computing adaptive learning rates (Bruno et al. 2021) for each parameter. Other variants have been widely used and have demonstrated their efficiency like AdaDelta (Sutskever et al. 2013) and Adamax (Kingma and Ba 2014). High-order optimization (Shanno 1970; Pajarinen et al. 2019) attracts widespread attention, but face more challenges. These methods are particularly useful where the objective function is highly non-linear and poorly conditioned. Newton-type methods use curvature information in the form of the Hessian matrix, in addition to the gradient. They are mainly introduced to extend high-order methods to large-scale data (Bollapragada et al. 2019). However, this family of methods has not been widely used in DL because of high per-iteration costs to store the inverse Hessian matrix. Other Newton-based methods have also been developed in the optimization literature (Byrd et al. 2016). When the derivative of the objective function is not easy to calculate, gradient-free techniques can be used (Berahas et al. 2019). In this sense, heuristic and metaheuristic techniques have been widely used. The recent literature involves numerous works using such techniques for ANN training such as Particle Swarm Optimization (PSO) (Shi 2004), Genetic Algorithm (GA) (Whitley et al. 1990), Improved whale trainer (IWT) (Khishe and Mosavi 2019), Chimp Optimization Algorithm (ChOA) (Jia et al. 2021), salp swarm algorithm (SSA) (Khishe and Mohammadi 2019), Adaptive Best-Mass Gravitational Search Algorithm (ABGSA) (Mosavi et al. 2019), Dragonfly Algorithm (DA) (Khishe and Safari 2019), The Arithmetic Optimization Algorithm (AOA) (Abualigah et al. 2021). However, the implementation strategy of heuristic and metaheuristic algorithms in large-scale deep learning problems is still rarely investigated (Berahas et al. 2019). Indeed, although metaheuristic techniques may provide satisfactory solutions in a reasonable time, their main limitation lies in the difficulty to handle high-dimensional and complex optimization problems, as well as convergence guarantees and stability. For specific cases, federated optimization has also been investigated in the recent literature developing Federated Learning (FL) techniques (Konečnỳ et al. 2016; Li et al. 2020; Yurochkin et al. 2019). Individual nodes hold a portion of the data, and the goal is to create a single common model that fits the entire distribution. A small batch gradient descent is generally used for weights optimization. FL is mainly useful when data portions can/must be kept locally on collaborating nodes. Specific variants such as fuzzy consensus have also been proposed (Połap 2021). Several approaches are investigated in the literature to solve the optimization issue for Machine Learning field. We summarize some of the mentioned optimization methods in terms of years of publications, purpose, advantages and disadvantages in Table 1. As described, the optimization approaches could be belong to some classes such as high-order methods, and metaheuristic methods.
Table 1

Summary of optimization methods

CategoriesMethodYearsPurposeAdvantagesDisadvantages
SGD Robbins and Monro (1951) and Sutskever et al. (2013)2013The update parameters are calculated using a randomly sampled mini-batch. The method converges at a sublinear rateThe computational time for each update does not depend on the total number of training samplesDifficult setting of an appropriate learning rate. The solution may be trapped at the saddle point in some cases
First-OrderAdam Kingma and Ba (2014)2014To dynamically adjust the learning rate of each parameter, use gradient first and second order moment estimationsStable gradient descent process. Suitable for most non-convex optimization problems with large data sets and high dimensional spaceThe method may not converge in some cases
Adadelta Zeiler (2012)2012Change the way of total gradient accumulation to exponential moving averageImproves the ineffective learning problem in the late stage of AdaGrad. Suitable for optimizing non-stationary and non-convex problemsIn the late training stage, the update process may be repeated around the local minimum
Adamax Kingma and Ba (2014)2017Generalization of Adam. It is based on adaptive lower-order moment estimationInfinite-order norm makes the algorithm stableThe penalty parameter is related to both the original and dual residuals whose value is difficult to determine
Newton’s method Avriel (2003)2003Calculates the inverse of the Hessian matrix to obtain faster convergence than with first-order approachesFaster convergence than the first-order gradient method. Quadratic convergence under certain conditionsLong computing time and large storage space at each iteration
High-orderQuasi-Newton method Nocedal and Wright (2006)2006Uses a Hessian matrix approximation or its inverseThere is no need to calculate the Hessian matrix’s inverse matrix, which reduces the computing time. It is possible to obtain superlinear convergence in most casesLarge storage space: not suitable for large-scale problems
Hessian free (HF) method Martens (2010)2010Sub-optimization with the conjugate gradient: avoids the computation of inverse Hessian matrixThe second-order gradient information can be used. There’s no need to calculate Hessian matrices directly. Suitable for high dimensional optimizationComputation cost to calculate the matrix–vector product increases linearly with the training data. Not appropriate for large-scale issues
Sochastic Quasi-Newton method Bottou et al. (2018)2018Employs techniques of stochastic optimization: e.g. online-LBFGS Schraudolph et al. (2007) and SQN Byrd et al. (2016)Can handle large-scale issuesMore complex that the stochastic gradient method
IWT Khishe and Mosavi (2019)2019Using a suitable spiral shape inspired by a humpback whale to improve the exploitation phase of the standard whale optimization algorithmStronger global search ability. It can be used to effectively solve complex constrained optimization problemsSlow convergence and easy to fall into local optimum
Derivative-free“Meta-heuristic”SSA Khishe and Mohammadi (2019)2019Is a bio-inspired optimization algorithm based on swarming mechanism of salps to enhance accuracy and reliability of the solutionFaster to execute because of its lower complexity. Improved capability in avoiding local minimaIt may get stuck in the local area, which results in the failure to obtain the global optimal solution
DA Khishe and Safari (2019)2019Inspired by the dynamic and static swarming behaviors of dragonflies to resolves local optima stagnation when solving challenging problemsSimple and easy to implement. Having few parameters for tuningIt does not have an internal memory that can lead to premature convergence to the local optimum
ABGSA Mosavi et al. (2019)2019Used to solve the problem of impertinent classification accuracy, and to block local minima as well as low convergence speed for Multi-Layer Perceptron Neural NetworkReduced complexity and processing timeUnaffordable sampling rate. Difficult because of randomness. Greatly influenced by initial solution
Federated optimizationFuzzy consensus Połap (2021)2021The goal of FL tasks is to learn a single global model that minimizes the empirical risk function over the entire training dataset. The authors extend FL using a fuzzy consensus method to improve large-scale group decision-making (LSGDM).In practical use, it has a great advantage due to the possibility of quick implementation and classification of the sample even during the training process. Capable of providing the most effective solution to complex issuesAdapting centralized training workflows such as hyperparameter tuning, and interpretability tasks to the federated learning setting present roadblocks to the widespread adoption of FL in practical settings
Moreover, the main advantages are mainly summarized in term of convergence speed, adaptation to high-dimensional optimization, precision, and prevention of local minima. Despite the existence of some works that may be suitable for non-convex problems, they should be improved especially for large-scale problems. Indeed, one can easily notice the slower convergence and the possibility of falling into a local optimum quite far from the global optimum, most of the optimization algorithms. On the other hand, the Bayesian framework has demonstrated its ability to provide reliable optimization models enjoying solid convergence guarantees and high stability level. Moreover, the flexibility of this framework allows introducing sophisticated constraints, such as those related to networks sparsification. As regards inference, MCMC-based techniques may also be adapted to large data problems (Quiroz et al. 2016). A Bayesian framework assumes that all parameters are realizations of random variables. Likelihood and prior distributions are formulated to model the available information on the target parameters. An estimator for these parameters is generally derived using a maximum a posteriori (MAP) framework. However, the main difficulty is to derive analytical closed-form expressions of the estimators because of the posterior distribution form which can be complex if sophisticated priors are used, such as those promoting sparsity. In this case, MCMC techniques are generally used to sample coefficients from the target posterior (Fakhfakh et al. 2020b). The main limitation of such techniques lies in the high complexity level, especially when multidimensional data are handled. In such cases, efficient sampling methods have been proposed in the literature such as the random walk Metropolis Hastings (MH) algorithm (Lee et al. 2012) or the Metropolis-adjusted Langevin algorithm (MALA) (Roberts and Tweedie 1996). Recently, sampling using Hamiltonian dynamics (Hanson 2001) has been investigated developing the so called Hamiltonian Monte Carlo (HMC) sampling. A more sophisticated algorithm has been proposed in Chaari et al. (2016) called non-smooth Hamiltonian Monte Carlo (ns-HMC) sampling. This method solves the problem of HMC schemes that cannot be used in the case of exponential distributions with non-differentiable energy function. The optimization methods in ANNs still face many challenges and open problems. There are mainly two major challenges with respect to data and model. The first one is insufficient training data, while the second is a non-convex objective function in DL architectures. In general, training a deep network requires large datasets to achieve good training. However, the lack of datasets to estimate the parameters in the learning models may lead to high variance (Chang et al. 2017) and overfitting (Hawkins 2004) problems. Regularization and Dropout are the most used techniques to alleviate the above-mentioned problems. In this paper, we investigate the use of ns-HMC for the learning process of ANNs. Specifically, we propose a Bayesian optimization method to minimize the target cost function and derive the optimal weights vector. The proposed method targets regularization schemes promoting sparse networks (Mocanu et al. 2018). Indeed, gradient-based optimization methods in this case are not very efficient due to differentiability and convergence issues. Learning performances can therefore be altered. We demonstrate that using the proposed method leads to high accuracy results with different CNN architectures, which cannot be reached using competing optimizers.

Problem formulation

It is well known that weights optimization is one of the key steps to design an efficient artificial neural network. For instance, if we consider a classification problem, the ANN weight vector W is updated during the learning phase by minimizing an error between the ground truth and the labels estimated using the network. An iterative procedure is generally performed, and gradient-based optimization procedures are used. For the sake of efficiency, regularization can also be performed in order to have a more accurate weights configuration. The sparse optimization method can be used for various tasks to produce sparse solutions. The penalty added to the classification cost can be interpreted as a convexification of the penalty. In Han et al. (2015), weights with the smallest amplitude in pretrained networks are removed. Model sensitivity to weights can also be used Tartaglione et al. (2018); Gomez et al. (2019), where weights with a weak influence on the network output are pruned. The norm, which counts the number of non-zero elements, is the most intuitive form of sparse regularizers and can promote the sparsest solution. However, minimizing problem is usually NP-hard (Natarajan 1995). The norm is the most commonly used surrogate, which can be solved easily. When applied in DNNs, sparse regularizer is supposed to zero redundant weights and thus remove unnecessary connections. However, if one aims at promoting sparse networks, sparse regularizations should be used, which makes the use of gradient-based algorithms inefficient since the error to be minimized in this case is no longer differentiable. In this paper, we propose a method to allow weights optimization under non-smooth regularizations. Let us denote by x an input to be presented to the ANN. The estimated label will be denoted by as a non-linear function of the input x and the weights vector , while the ground truth label will be denoted by y. Using a quadratic error with an regularization with M input data for the learning step, the weights vector can be estimated aswhere is a regularization parameter balancing the solution between the data fidelity and regularization terms, and M is the number of learning data. It is worth noting that other regularization terms can be used in Eq. (1) (, ,...). The norm is used to promote weights sparsity. Summary of optimization methods Since the optimization problem in Eq. (1) is not differentiable, the use of gradient-based algorithms with back-propagation is not possible and the learning process is costly and very complicated. In Sect. 4 we present a method to efficiently estimate the weights vector without an increase of learning complexity. The optimization problem in Eq. (1) is formulated and solved in a Bayesian framework. This formulation has two main advantages. The first one is related to the flexibility of such models to handle a large panel of regularization terms through an exponential formulation to mimic the variational form in Eq. (1). The second advantage is related to the ability to design fully automatic schemes without user intervention/configuration. Indeed, this is very important, especially for complex problems where parameters fitting are complicated.

Bayesian optimization

As stated above, the weights optimization problem is formulated in a Bayesian framework. In this sense, the problem parameters and hyperparameters are assumed to follow probability distributions. More specifically, a likelihood distribution is defined to model the link between the target weights vector and the data, while a prior distribution is defined to model the prior knowledge about the target weights.

Hierarchical Bayesian model

According to the principle of minimizing the error between the reference label y and the estimated one , and assuming a quadratic error (first term in (1)), we define the likelihood distribution aswhere is a positive parameter to be set. As regards prior information about the target weights, and to promote sparsity of the estimated vector (and hence the sparsity of the deep network), a common choice is to resort to an penalization. Under a Bayesian framework, the Laplace distribution can be used.where is a hyperparameter to be set. This prior allows us to introduce exactly the same prior information as the norm in Eq. (1). By adopting a MAP approach, we first need to express the posterior distribution. Based on the defined likelihood and prior, this posterior writes:It is clear that this posterior is not straightforward to handle in order to derive a closed-form expression of the estimate . For this reason, we resort to a stochastic sampling approach in order to numerically approximate the posterior, and hence to calculate an estimator for . The following Section details the adopted sampling procedure.

Hamiltonian Monte Carlo

Hamiltonian Monte Carlo (Li et al. 2015) is a class of sampling algorithms inspired by the Hamiltonian dynamics. It is a reformulation of the theory of classical mechanics which is intended to describe the motion of objects, and therefore to model the dynamic physical systems (Alder and Wainwright 1959). A dynamic particle of mass m can be characterized essentially by its positions W and momentum , where which represent the velocity of the particle. The Hamiltonian system models the total Kinetic of thus particle, respectively, the potential energy E(W) and the Kinetic energy , which can likewise be expressed as a function of momentum by . Thus, the Hamiltonian H(W, q) can be expressed as:nd the dynamics of the particle can be specified by a set of coupled differential equations (Neal 2011),From any time interval of duration s, these equations define a mapping , from the state at any time t to the state at time . When estimating a random variable with probability density function in the HMC method, we define an auxiliary momentum variable q. The pdf of the Hamiltonian dynamic energy defined in (5) is given byHMC methods iteratively proceed with updating W and q by sampling according to the distribution (8). The sampling of the model is performed by two-step. The first sampling q according to the multivariate Gaussian distribution N(0,), where is the identity matrix. The second step, updates both momentum q and position W by proposing two candidates and . These two candidates are generated by simulating the Hamiltonian dynamics, which are discretized using some discretization techniques which is leapfrog method (Hanson 2001). The discretization can be performed using steps of the leapfrog method with a stepsize : can either be manually fixed or automatically tuned (Wang et al. 2013).

ns-HMC

As the HMC is a successful approach for sampling from continuous densities, however, it has difficulty simulating Hamiltonian dynamics with non-smooth functions, leading to poor performance. A novel scheme called Non-smooth Hamiltonian Monte Carlo (ns-HMC) has been proposed in Chaari et al. (2016) to make feasible the use of Hamiltonian dynamics of sampling even for target distributions with non-smooth energy functions. The sampling technique relies on some interesting results from convex optimization and Hamiltonian Monte Carlo methods. The main idea of the ns-HMC scheme is to modify the leapfrog discretization scheme by introducing a step calculation the proximity operator of . The detailed ns-HMC scheme is given by algorithm 1, where represent the number of leapfrog step and is the stepsize. However, analytic calculation of the proximity operator for a wide class of energy functions is not possible. This drawback prevents the use of the ns-HMC algorithm in the case of sparse target distributions where the proximity operator of the energy function is difficult to be calculated. In order to solve this problem, in Chaari et al. (2017), a modified ns-HMC sampling scheme, called general ns-HMC, have been proposed involving a Bayesian calculation of the proximity operator. Thus, instead of calculating the proximity operator at each step as shown in the Algorithm 1 that can be led to an increased computational cost, with the general ns-HMC scheme, the calculation of the proximity operator is only calculated at the initialization step. The calculated value is then used to update the proximity operator value at different points. Another advantage of the method is that it does not depend on the initial point where the proximity operator is calculated first.

Hamiltonian sampling

Let us denote , . For a weight we define the following energy functionThe posterior in (4) can be reformulated asTo sample according to this exponential posterior, and since direct sampling is not possible due to the form of the energy function , Hamiltonian sampling is adopted. Indeed, Hamiltonian dynamics (Hanson 2001) strategy has been widely used in the literature to sample high dimensional vectors. However, sampling using Hamiltonian dynamics requires computing the gradient of the energy function, which is not possible in our case due to the term. To overcome this difficulty, we resort to a non-smooth Hamiltonian Monte Carlo (ns-HMC) strategy. Indeed, this strategy requires calculation of the proximity operator only at an initial point and uses the shift property (Moreau 1965) to deduce the proximity operator during the iterative procedure. As regards the proximity operator calculation, let us denote by the gradient of the quadratic term of the loss function with respect to the weight . Let us also denote by . Following the proximity operator standard definition, we can write for any real zStraightforward calculations lead to the following expression of the proximity operator:Since is nothing but the soft thresholding operator (Chaux et al. 2007), the proximity operator in (11) can be easily calculated once a single gradient step is applied (back-propagation) to calculate . We therefore propose the following leapfrog discretization scheme to be integrated in Algorithm 1, where = :The Gibbs sampler resulting from the proposed leapfrog discretization scheme is summarized in Algorithm 2. The proposed candidates are given by and after leapfrog steps. These candidates are then accepted based on the standard MH rule, i.e., with the following probabilitywhere H is the Hamiltonian defined in (5). After convergence, Algorithm 2 provides chains of coefficients sampled according to the target distribution of each . These chains can be used to compute an MMSE (minimum mean square error) estimator after discarding the samples corresponding to the burn-in period. It is worth noting that hyperprior distributions can be put on and in order to integrate them in the hierarchical Bayesian model. These hyperparameters can therefore be estimated from the data at the expense of some additional complexity.

Experimental validation

In order to validate the proposed method, five image classification experiments are conducted using four datasets: two COVID-19 datasets including Computed tomography (CT) images for simple (Angelov and Almeida Soares 2020) and challenging classification (Yang et al. 2020), and two standard datasets, namely Fashion-MNIST (Xiao et al. 2017) and CIFAR-10 Recht et al. (2018). Table 2 illustrates the setting details of the different datasets.
Table 2

Setting details of the used datasets.

DatasetTraining setTest set# Classes
CT images for simple classification12104302
CT images for challenging classification5661802
Fashion-MNIST48,00012,00010
CIFAR-1050,00010,00010
Setting details of the used datasets. In order to compare the proposed method with the state of the art, three kinds of optimizers are used : (i) MCMC-based methods, precisely the standard Metropolis-Hastings (MH) algorithm (Chib and Greenberg 1995) and its random walk variant (rw-MH), (ii) the most popular and widely used techniques : Adam, Adamax, SGD , Adadelta, and (iii) three metaheuristics algorithms: Improved Whale Trainer, Dragonfly Algorithm, and Salp Swarm Algorithm. The parameters setting of all these algorithms is detailed in Table 3.
Table 3

Parameters setting for benchmark algorithms

AlgorithmParametersDescriptionValues
MH\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sigma$$\end{document}σStand. dev. of the proposal normal distribution3
rw-MH\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\sigma$$\end{document}σStand. dev. of the proposal normal distribution5
lrLearning rate\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$10^{-3}$$\end{document}10-3
Adam\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta _1$$\end{document}β11st moment estimates exponential decay rate0.9
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta _2$$\end{document}β22nd moment estimates exponential decay rate0.999
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\epsilon$$\end{document}ϵNumerical stability constant1e\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$-$$\end{document}-08
lrLearning rate\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$10^{-3}$$\end{document}10-3
SGDmomentumAcceleration rate0.8
decayLearning rate decay over each update1e\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$-$$\end{document}-6
lrLearning rate\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$10^{-3}$$\end{document}10-3
AdadeltarhoDecay rate0.95
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\epsilon$$\end{document}ϵNumerical stability constant1e\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$-$$\end{document}-08
lrLearning rate\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$10^{-3}$$\end{document}10-3
Adamax\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta _1$$\end{document}β11st moment estimates exponential decay rate0.9
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\beta _2$$\end{document}β22nd moment estimates exponential decay rate0.999
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\epsilon$$\end{document}ϵNumerical stability constant1e\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$-$$\end{document}-08
minvLower bound\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$-$$\end{document}- 2
IWTmaxvUpper bound2
sizeNumber of particles30
p_sSpiral parameter3
minvLower bound\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$-$$\end{document}- 2
DAmaxvUpper bound2
sizeNumber of particles15
minvLower bound\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$-$$\end{document}- 2
SSAmaxvUpper bound2
sizeNumber of particles30
As regards coding, we used python programming language with Keras and Tensorflow libraries on an Intel(R) Core(TM) i7-2720QM CPU 2.20GHZ architecture with 16 Go memory. Parameters setting for benchmark algorithms

ConvNet models

Three CNN architectures are used in this study. Like the LeNet model (LeCun et al. 1998), the first one includes three convolutional (Conv-32,Conv-64 and Conv-128) and two fully-connected (FC-64 and FC-softmax). The second one has six convolutional (2XConv-32, 2XConv-64 and 2XConv-128) and three FC layers (FC-128, FC-64 and FC-softmax) that are organized similarly to VGG-Net (Muhammad et al. 2018). These architectures are shown in Table 4. The third one is a more deeper CNN used for comparison includes 15 convolutional layers and a FC-softmax layer (see Sect. 5.7 for more details).
Table 4

Convnet with regularization techniques

CNN_1CNN_2
Conv3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 3-32:stride=1             Conv3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 3-32:stride=1
BatchNormalization             BatchNormalization
MaxPool 2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 2             Conv3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 3-32:stride=1
Dropout(0.2)             BatchNormalization
      MaxPool 2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 2
      Dropout(0.2)
Conv3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 3-64:stride=1             Conv3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 3-64:stride=1
BatchNormalization             BatchNormalization
MaxPool 2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 2             Conv3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 3-64:stride=1
Dropout(0.3)             BatchNormalization
      MaxPool 2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 2
      Dropout(0.3)
Conv3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 3-128:stride=1             Conv3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 3-128:stride=1
BatchNormalization             BatchNormalization
MaxPool 2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 2             Conv3 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 3-128:stride=1
Dropout(0.4)             BatchNormalization
      MaxPool 2 \documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\times$$\end{document}× 2
      Dropout(0.4)
Flattening             Flattening
FC-64             FC-128
Dropout(0.3)             Dropout(0.3)
      FC-64
      Dropout(0.2)
FC-softmax             FC-softmax
All of them involve convolutional layers with Kernel filters in addition to max-pooling, with stride size equal to 1. All layers in the different configurations used ReLU as an activation function except the output layer. As deep neural networks can easily overfit when trained with small datasets, the used CNNs are extended with three regularizing techniques: Batch normalization (Ioffe and Szegedy 2015): deals with the change of the feature space distribution along with the model during the training. The input of the layer is normalized to be zero-mean with unitary variance. This step not only acts as a regularizer, but also allows for faster training, higher learning rates, and less dependence on weights initialization. Regularization (Xu et al. 2010): regularization is the preferred choice when having a high number of features as it provides sparse solutions. It allows obtaining the computational advantage because features with zero coefficients can be avoided. In our case, the used regularization parameter was set to . Dropout (Srivastava et al. 2014): random disabling of neurons during training with probability (or percentage) p. Temporarily ignoring some activations forces the other neurons to learn a more robust representation of the input data while reducing the sensitivity of specific neurons. In our study, the dropout rate is set by cross validation to . Convnet with regularization techniques

Sparsity and stability analysis

In this section, we evaluate the sparsity and the stability of the estimation of the weights with different values of and compare to Adam as a state of the art optimizer. The CNN_1 architecture is applied using the CT Covid-19 image database. Table 5 reports accuracy, computational time, and norm of the estimated weights using different values of the regularization parameter over 10 Monte Carlo runs. To further evaluate the sparsity level, Table 5 also reports the pseudo-norm values (number of non-zeros). Standard deviations over the 10 runs are provided in the table.
Table 5

Accuracy, sensitivity, specificity, computational time (in min), and norms of the estimated weights for CNN_1 using Adam and the proposed method with different values of

Optimizer\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\lambda$$\end{document}λ\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\Vert \cdot \Vert _0$$\end{document}·0\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\Vert \cdot \Vert _1$$\end{document}·1Acc.TimeSens.Spec.
ns-HMC\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$10^{-3}$$\end{document}10-3149,11348,98889.6837.2888.2186.95
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 9.07$$\end{document}±9.07\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 9.10$$\end{document}±9.10\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.04$$\end{document}±0.04\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.63$$\end{document}±0.63\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.06$$\end{document}±0.06\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.08$$\end{document}±0.08
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$10^{-2}$$\end{document}10-2142,54245,47690.0238.7989.0288.57
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm \mathbf{8}.78$$\end{document}±8.78\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm \mathbf{8}.81$$\end{document}±8.81\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm \mathbf{0}.02$$\end{document}±0.02\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm \mathbf{0}.61$$\end{document}±0.61\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm \mathbf{0}.4$$\end{document}±0.4\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm \mathbf{0}.3$$\end{document}±0.3
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$10^{-1}$$\end{document}10-1152,90451,23289.4238.4787.8186.11
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 9.11$$\end{document}±9.11\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 9.19$$\end{document}±9.19\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.05$$\end{document}±0.05\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.70$$\end{document}±0.70\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.5$$\end{document}±0.5\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.6$$\end{document}±0.6
Adam\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$10^{-3}$$\end{document}10-3180,51351,72786.9152.1284.3681.25
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 9.77$$\end{document}±9.77\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 9.11$$\end{document}±9.11\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.07$$\end{document}±0.07\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.87$$\end{document}±0.87\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.9$$\end{document}±0.9\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.8$$\end{document}±0.8
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$10^{-2}$$\end{document}10-2191,22967,73285.3454.9183.1480.66
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 10.28$$\end{document}±10.28\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 10.63$$\end{document}±10.63\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.12$$\end{document}±0.12\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 1.08$$\end{document}±1.08\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 1.12$$\end{document}±1.12\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 1.09$$\end{document}±1.09
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$10^{-1}$$\end{document}10-1189,07558,82385.4953.8584.0982.49
\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 10.15$$\end{document}±10.15\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 10.47$$\end{document}±10.47\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.09$$\end{document}±0.09\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 0.95$$\end{document}±0.95\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 1.05$$\end{document}±1.05\documentclass[12pt]{minimal} \usepackage{amsmath} \usepackage{wasysym} \usepackage{amsfonts} \usepackage{amssymb} \usepackage{amsbsy} \usepackage{mathrsfs} \usepackage{upgreek} \setlength{\oddsidemargin}{-69pt} \begin{document}$$\pm 1.01$$\end{document}±1.01

Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

The obtained scores clearly indicate that our method provides estimates with a higher sparsity level in comparison to Adam: the number of non-zero weights is significantly lower than the Adam optimizer. The proposed method provides estimates with 14% higher sparsity level. Moreover, the reported low standard variation values in Table 5 clearly indicate good stability properties of the proposed method with respect to random sampling in the MCMC procedure, which confirms the good convergence properties. This stability holds for accuracy, sparsity, sensitivity, specificity, and computational time. Moreover, the same conclusions hold for all tested values. In other words, the same conclusions hold for different network sparsity levels. Accuracy, sensitivity, specificity, computational time (in min), and norms of the estimated weights for CNN_1 using Adam and the proposed method with different values of Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

Experiment 1: COVID-19 classification using CT images

This section studies the performance of our optimizer for classifying CT data into normal and Covid-19 cases using a public dataset of CT scans for SARS-CoV-2 identification. The dataset is made up of of 1252 CT scans of size that are positive for SARS-CoV-2 infection (COVID-19) and 1230 CT scans for negative patients. These data have been collected from real patients in hospitals from Sāo Paulo, Brazil.1 Table 6 reports accuracy, loss, sensitivity, specificity and computational time, for all optimizers with CNN_1 and CNN_2. The reported scores indicate that our ns-HMC outperforms the competing optimizers, including metaheuristic methods in terms of learning precision, and hence classification performances. Accuracy values even show a slight advantage in favor of the proposed method for both the CNNs. The lack of performance obtained by the metaheuristic methods (IWT, DA, and SSA) caused by early convergence experienced during the process of finding optimum value. This phenomenon is due to the lack of population diversity and known as premature convergence. The proposed method enjoys faster convergence properties while reaching global optimum due to the Bayesian formulation with lower loss rates.
Table 6

Experiment 1: Results for CT image classification using CNN_1 and CNN_2 (Computational time in min, accuracy, loss, sensitivity and specificity)

OptimizersCNN_1CNN_2
Time (min)Acc.LossSens.Spec.Time (min)Acc.LossSens.Spec.
ns-HMC370.900.100.890.87580.920.090.900.89
MH79.20.840.180.810.77133.80.860.160.850.80
rw-MH64.80.850.170.840.7995.40.870.140.860.82
Adam520.870.120.860.8385.20.880.110.870.85
SGD530.880.130.850.8087.10.860.150.840.81
Adadelta560.860.120.840.8190.60.870.110.850.83
Adamax530.870.130.860.8487,60.870.120.860.85
IWT590.840.210.810.7870,20.860.190.840.83
DA610.830.250.820.7973.20.850.220.830.81
SSA570.860.180.850.8361.20.880.170.870.86

Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

Experiment 1: Results for CT image classification using CNN_1 and CNN_2 (Computational time in min, accuracy, loss, sensitivity and specificity) Bold values indicate the importance of the obtained results of our approach compared to competing algorithms The behavior of the algorithms during the training step is displayed in Figs. 1, and 2 where the curves clearly indicate a convergence with high accuracy rate for most optimizers. A significant difference between training and loss curves may indicate potential overfitting obtained with some optimizers. This gap is reduced using the proposed optimizer. Interestingly, the same behavior is observed for both CNN models. Moreover, the accuracy increase between CNN_1 and CNN_2 is almost the same for all optimizers (see Table 6). The higher performance of the proposed method can be explained by a better exploration of the searching space due to the Bayesian formulation and the efficient sampling scheme, which also helps reducing the computational time. Indeed, ns-HMC sampling integrates a gradient information related to the geometry of the target distribution, which finally leads to a faster convergence of the used sampler.
Fig. 1

Experiment 1: Train and test curves using CNN_1

Fig. 2

Experiment 1: Train and test curves using CNN_2

It is worth noting that the curves irregularity for Bayesian techniques (proposed method, MH and rw-MH) are due to the random sampling effect. No monotonic behavior is expected. Experiment 1: Train and test curves using CNN_1 Experiment 1: Train and test curves using CNN_2

Experiment 2: challenging case

A more challenging classification case is addressed in this experiment. The same CNNs are used for CT images classification to identify COVID-19 infections from other pneumonia. In contrast to Experiment 1, this task is challenging due to the rich content of CT images and similarity between COVID-19 infection and other pneumonia. The COVID-CT dataset contains 349 CT images positive for COVID-19 belonging to 216 patients and 397 CT images that are negative for COVID-19. The dataset is open-sourced to the public2. We used 566 images for the train and 180 images for the test with size of . The reported scores in Table 7 indicate that the proposed method clearly outperforms the competing optimizers in both models to solve this challenging classification problem. Moreover, severe performance decrease is observed for some optimizers. IWT, DA, and SSA achieved an accuracy slightly better than gradient and MCMC-based methods. DA algorithm is the better performer compared to all the competing algorithms on this dataset, but has an accuracy less than our nc-HMC optimizer of around 6%. This is mainly due to this challenging classification, which leads to a more complex learning process.
Table 7

Experiment 2: Results for CT image classification - challenging case - using CNN_1 and CNN_2 (computational time in min, accuracy, loss, sensitivity and specificity)

OptimizersCNN_1CNN_2
Time (min)Acc.LossSens.Spec.Time (min)Acc.LossSens.Spec.
ns-HMC400.840.260.820.80530.880.220.860.85
MH71,40.730.380.710.6992.40.760.340.740.72
rw-MH590.760.360.750.7294.80.770.320.750.74
Adam580.710.430.690.68810.730.360.720.71
SGD590.650.450.640.6282.20.680.420.670.65
Adadelta61.80.670.420.650.6387.60.700.380.690.67
Adamax60.60.690.410.670.66900.740.360.720.71
IWT540.750.380.740.72900.780.350.770.75
DA570.780.360.770.76870.810.330.800.76
SSA510.760.370.760.75830.790.360.780.77

Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

Experiment 2: Results for CT image classification - challenging case - using CNN_1 and CNN_2 (computational time in min, accuracy, loss, sensitivity and specificity) Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

Experiment 3: Fashion-MNIST image classification

In this scenario, the learning performance using the competing optimization algorithms is evaluated using the standard Fashion-MNIST dataset. A training set of 60,000 images is used, while the test set was made up of 10,000 images. Each example is a grayscale image, associated with a label from 10 classes, with 7000 images per class. For the model training, we used 48,000 images for the train set and 12,000 for the test. The obtained results for the fashion-MNIST dataset is given in Table 8. All the competing optimizers did not perform well on this dataset which could be because of the size of the dataset. Our optimizer was the better performer with an accuracy up to 93% for both architectures, significantly better than all the competing optimizers. Indeed, as reported in Table 8, the computational time of all the competing algorithms is generally around 160 min for the CNN_1, more than twice the time needed for the proposed method. The same conclusion for the deep architecture CNN_2.
Table 8

Experiment 3: Results for Fashion-MNIST image classification—challenging case—using CNN_1 and CNN_2 (computational time in min, accuracy, loss, sensitivity and specificity)

OptimizersCNN_1CNN_2
Time (min)Acc.LossSens.Spec.Time (min)Acc.LossSens.Spec.
ns-HMC70.50.920.229088308.40.930.199189
MH166.20.860.350.850.81745.80.870.330.840.82
rw-MH183.60.880.330.860.83797.40.880.310.850.84
Adam156.60.900.460.850.824440.920.320.880.87
SGD164.40.880.710.710.67452.40.890.560.840.83
Adadelta169.80.701.200.660.64439.80.780.960.710.70
Adamax1490.910.490.880.82448.20.910.260.880.87
IWT180.20.820.370.790.734860.830.360.800.79
DA1740.790.400.760.744690.820.350.780.78
SSA165.70.840.330.810.754530.860.240.860.85

Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

Experiment 3: Results for Fashion-MNIST image classification—challenging case—using CNN_1 and CNN_2 (computational time in min, accuracy, loss, sensitivity and specificity) Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

Experiment 4: CIFAR-10 image classification

In this scenario, the learning performance using the competing optimization algorithms is evaluated using the standard CIFAR-10 dataset. The CIFAR-10 dataset consists of 60,000 32 32 color images in 10 classes, with 6000 images per class. There are 50,000 training images and 10,000 test images. The dataset is divided into five training batches and one test batch, each with 10,000 images. The test batch contains exactly 1000 randomly-selected images from each class. The training batches contain the remaining images in random order, but some training batches may contain more images from one class than another. Between them, the training batches contain exactly 5000 images from each class. Classification results for the CIFAR-10 dataset are given in Table 9.
Table 9

Experiment 4: Results for CIFAR-10 image classification—challenging case—using CNN_1 and CNN_2 (computational time in min, accuracy, loss, sensitivity and specificity)

OptimizersCNN_1CNN_2
Time (min)Acc.LossSens.Spec.Time (min)Acc.LossSens.Spec.
ns-HMC85.70.910.2589873310.920.219087
MH1720.830.410.800.78763.20.840.360.830.81
rw-MH192.20.850.360.840.81814.10.860.350.840.83
Adam1610.890.420.830.814290.900.360.870.86
SGD1690.860.750.690.65459.70.860.600.830.80
Adadelta174.70.750.920.680.65453.30.790.810.720.70
Adamax1550.900.330.880.85507.80.910.240.870.85
IWT1860.800.350.780.745310.810.340.790.78
DA1790.820.350.770.755190.840.300.780.76
SSA172.70.830.310.810.775030.850.270.830.80

Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

The same conclusion can be drawn as that of the Fashion-MNIST dataset. The proposed Bayesian optimizer showed great overall performance even if more classes are considered compared to all competing optimizers. Experiment 4: Results for CIFAR-10 image classification—challenging case—using CNN_1 and CNN_2 (computational time in min, accuracy, loss, sensitivity and specificity) Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

Comparison on deep CNN

This section studies the performances of our ns-HMC method using a deep CNN on the standard Fashion-MNIST dataset. The proposed deep CNN is deeper than and . It is made up of 15 convolutional layers (5 Conv3 3-32, 5 Conv3 3-64, 5 Conv3 3-128) and a FC-softmax layer. All of them involve convolutional layers with 3 3 Kernel filters in addition to 2 2 max-pooling, with stride size equal to 1. The use of a deep CNN validates the effectiveness and robustness of our approach in terms of accuracy, loss, sensitivity, and specificity criteria as shown in Table 10. Furthermore, most of the competing optimizers clearly indicate an overfitting effect like SGD, and Adadelta, in contrast to the proposed method. Hence, one can easily notice the highest global accuracy of our Bayesian optimizer regardless of the depth architectures.
Table 10

Results for fashion-MNIST image classification using deep CNN (computational time in min, accuracy, loss, sensitivity and specificity)

OptimizersTime (min)Acc.LossSens.Spec.
ns-HMC5820.930.200.910.91
MH9770.840.410.800.76
rw-MH9860.850.370.830.78
Adam7010.910.550.880.85
SGD7050.880.460.820.79
Adadelta7070.800.630.700.72
Adamax7060.920.440.900.88
IWT6810.880.330.820.79
DA6770.850.370.810.77
SSA6940.900.310.860.84

Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

Results for fashion-MNIST image classification using deep CNN (computational time in min, accuracy, loss, sensitivity and specificity) Bold values indicate the importance of the obtained results of our approach compared to competing algorithms

Discussion

In this paper, we proposed a novel optimization method built in a Bayesian framework. The proposed algorithm relies on a Hamiltonian Monte Carlo scheme to solve the subsequent optimization problem involving sparsity constraints, while being adapted to large data problems under solid convergence guarantees. The proposed method has been validated on four different datasets in order to assess: its (i) efficiency on a classification case where competing optimizers provide good results, (ii) fast convergence properties, and (iii) robustness with respect to the sample size. A gold standard validation on the widely used both Fashion-MNIST and CIFAR-10 databases have also been performed. Furthermore, three kinds of comparisons are performed: (i) with respect to other state of the art optimizers (Adam, SGD and Adadelta), (ii) with respect to other MCMC-based techniques (MH and rw-MH), and (iii) with respect to the novel metaheuristics methods (IWT, DA, and SSA). Though deep neural network has good expressive ability, its large model parameters which bring a great burden on calculation is still a problem remain to be solved. This problem hinders the widely deployment of DNNs-based applications, so it is primary of reducing the model parameters without losing performance. Sparsing neural networks is one of the methods to effectively reduce complexity which can improve efficiency and generalizability. Empirical evidence shows that deep architectures often require to be over-parametrized (having more parameters than training examples) in order to be successfully trained (Brutzkus et al. 2017; Mhaskar and Poggio 2016). Indeed, the use of these networks is useful to extract more implicit characteristics that leads to good precision of the model, and hence reduce the overfitting effect. However, once input-output relations are properly represented by a complex network, such a network may form a starting point in order to find a simpler, sparser, but sufficient architecture (Brutzkus et al. 2017; Mhaskar and Poggio 2016). Three CNNs which one is deeper than others have been used with regularization to promote sparse networks. This allowed us to analyse how the proposed optimizer behaves when the complexity level of the network increases. From one side, experiments showed that our optimizer enjoys better sparsity levels in terms of and norms. Experiments lead one to conclude that the proposed non-smooth Hamiltonian sampling scheme provides faster and more accurate convergence with lower overfitting effects. The obtained gain is not only due to the Bayesian formulation, but also to the efficient inference scheme. From another side, results showed that better accuracy is always obtained with the most sophisticated CNN in spite of the additional complexity. The proposed method speeds up the learning time for both architectures. Indeed, for the investigated challenging classification problem (Experiment 2), a deeper network did not solve the overfitting problem with standard optimizers, while our method overcomes this limitation by providing more accurate optimization of the target criterion, and only need simple neural net architecture to produce high accuracy. The sensitivity and specificity scores justify the stable behavior of our ns-HMC at different used datasets. Moreover, loss and accuracy obtained with our method do not fall when data complexity increases, in contrast to other competing optimizers. Metaheuristic methods have attracted considerable attention in the last years, mainly due to their simple heuristics and ability to optimize nondifferentiable functions. Indeed, the performance of a metaheuristic can only be examined in a problem where these are applicable. They do not guarantee a global optimum, but it is near to a global best solution. We can further conclude from the results that our optimization strategy is insensitive to local minima unlike to metaheuristic which are already rarely conducted optimize DL method (Rere et al. 2016). The main limitation of our proposed method concerns that is trained on CPU and not GPU which is commonly used for deep learning. Moreover, is a hyperparameter to be set as a fixed value in our method. The choice of the best value of depends on the used model for a training dataset. One of the challenges in the future is to extend our optimizer by estimating this hyperparameter.

Conclusion

In this paper, we proposed a new Bayesian optimization method to fit weights for sparse artificial neural networks. The proposed method relies on Hamiltonian dynamics with non-smooth regularizations, using a plug and play procedure. The proposed ns-HMC optimizer showed promising results with good classification performances and high generalization properties, in addition to low computational time in comparison with all the competing algorithms including the commonly used optimizers in DL and the recent metaheuristics methods. The use of standard datasets (Fashion-MNIST and CIFAR-10) with multiple images confirms the stability of our optimizer in terms of accuracy. The different experiments carried out have showed the good generalization of our ns-HMC and their ability to be applied in various fields. We can extend our experiments in the future by integrating the segmentation methods in Deep Learning by using our optimizer. Moreover, We will focus on investigating parallel implementation of the proposed method to further decrease computational time. Investigating the use of the proposed method on recurrent networks will also be considered.
  6 in total

1.  The problem of overfitting.

Authors:  Douglas M Hawkins
Journal:  J Chem Inf Comput Sci       Date:  2004 Jan-Feb

2.  A Survey of Optimization Methods From a Machine Learning Perspective.

Authors:  Shiliang Sun; Zehui Cao; Han Zhu; Jing Zhao
Journal:  IEEE Trans Cybern       Date:  2019-11-18       Impact factor: 11.448

3.  Detection and classification of lung diseases for pneumonia and Covid-19 using machine and deep learning techniques.

Authors:  Shimpy Goyal; Rajiv Singh
Journal:  J Ambient Intell Humaniz Comput       Date:  2021-09-18

4.  Scalable training of artificial neural networks with adaptive sparse connectivity inspired by network science.

Authors:  Decebal Constantin Mocanu; Elena Mocanu; Peter Stone; Phuong H Nguyen; Madeleine Gibescu; Antonio Liotta
Journal:  Nat Commun       Date:  2018-06-19       Impact factor: 14.919

5.  Metaheuristic Algorithms for Convolution Neural Network.

Authors:  L M Rasdi Rere; Mohamad Ivan Fanany; Aniati Murni Arymurthy
Journal:  Comput Intell Neurosci       Date:  2016-06-08

6.  Convolutional Neural Network for Drowsiness Detection Using EEG Signals.

Authors:  Siwar Chaabene; Bassem Bouaziz; Amal Boudaya; Anita Hökelmann; Achraf Ammar; Lotfi Chaari
Journal:  Sensors (Basel)       Date:  2021-03-03       Impact factor: 3.576

  6 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.