Literature DB >> 25404940

An effective hybrid cuckoo search algorithm with improved shuffled frog leaping algorithm for 0-1 knapsack problems.

Yanhong Feng1, Gai-Ge Wang2, Qingjiang Feng3, Xiang-Jun Zhao2.   

Abstract

An effective hybrid cuckoo search algorithm (CS) with improved shuffled frog-leaping algorithm (ISFLA) is put forward for solving 0-1 knapsack problem. First of all, with the framework of SFLA, an improved frog-leap operator is designed with the effect of the global optimal information on the frog leaping and information exchange between frog individuals combined with genetic mutation with a small probability. Subsequently, in order to improve the convergence speed and enhance the exploitation ability, a novel CS model is proposed with considering the specific advantages of Lévy flights and frog-leap operator. Furthermore, the greedy transform method is used to repair the infeasible solution and optimize the feasible solution. Finally, numerical simulations are carried out on six different types of 0-1 knapsack instances, and the comparative results have shown the effectiveness of the proposed algorithm and its ability to achieve good quality solutions, which outperforms the binary cuckoo search, the binary differential evolution, and the genetic algorithm.

Entities:  

Mesh:

Year:  2014        PMID: 25404940      PMCID: PMC4227462          DOI: 10.1155/2014/857254

Source DB:  PubMed          Journal:  Comput Intell Neurosci


1. Introduction

The application of nature-inspired metaheuristic algorithms to computational optimization is a growing trend [1]. Many hugely popular algorithms, including differential evolution (DE) [2, 3], harmony search (HS) [4, 5], krill herd algorithm (KH) [6-13], animal migration optimization (AMO) [14], grey wolf optimizer (GWO) [15], biogeography-based optimization (BBO) [16, 17], gravitational search algorithm (GSA) [18], and bat algorithm (BA) [19, 20], perform powerfully and efficiently in solving diverse optimization problems. Many metaheuristic algorithms have been applied to solve knapsack problems, such as evolutionary algorithms (EA) [21], HS [22], chemical reaction optimization (CRO) [23], cuckoo search (CS) [24-26], and shuffled frog-leaping algorithm (SFLA) [27]. To better understand swarm intelligence please refer to [28]. In 2003, Eusuff and Lansey firstly proposed a novel metaheuristic optimization method: SFLA, which mimics a group of frogs to search for the location that has the maximum amount of available food. Due to the distinguished benefit of its fast convergence speed, SFLA has been successfully applied to handle many complicated optimization problems, such as water resource distribution [29], function optimization [30], and resource-constrained project scheduling problem [31]. CS, a nature-inspired metaheuristic algorithm, is originally proposed by Yang and Deb in 2009 [32], which showed some promising efficiency for global optimization. Owing to the outstanding characteristics such as fewer parameters, easy implementation, and rapid convergence, it is becoming a new research hotspot in swarm intelligence. Gandomi et al. [33] first verified structural engineering optimization problems with CS algorithm. Walton et al. [34] proposed an improved cuckoo search algorithm which involved the addition of information exchange between the best solutions and tested their performance with a set of benchmark functions. Recently, the hybrid algorithms that combined CS with other methods have been proposed and have become a hot topic studied by people, such as CS combined with a fuzzy system [35], a DE [36], wind driven optimization (WDO) [37], artificial neural network (ANN) [38], and genetic algorithm (GA) [39]. For details, see [40]. In 2011, Layeb [25] developed a variant of cuckoo search in combination with quantum-based approach to solve knapsack problems efficiently. Subsequently, Gherboudj et al. [24] utilized purely binary cuckoo search to tackle knapsack problems. A few scholars consider binary-coded CS and its performance need to further improve so as to further expand its fields of application. In addition, despite successful application to the solution of 0-1 knapsack problem by many methods, in fact, it is still a very active research area, because many existing algorithms do not cope well with some new and more intractable 0-1 knapsack problems hidden in the real world. Further, most of recently proposed algorithms focused on solving 0-1 knapsack problems with low dimension and medium dimension, but 0-1 knapsack problems with high dimension are involved little and the results are not highly satisfactory. What is more, the correlation between the weight and the value of the items may not be more concerned. This necessitates new techniques to be developed. Therefore, in this work, we propose a hybrid CS algorithm with improved SFLA (CSISFLA) for solving 0-1 knapsack problem. To verify effectiveness of our proposed method, a large number of experiments on 0-1 knapsack problem are conducted and the experimental results show that the proposed hybrid metaheuristic method can reach the required optima more effectively than CS, DE, and GA even in some cases when the problem to be solved is too complicated and complex. The rest of the paper is organized as follows. Section 2 introduces the preliminary knowledge of CS, SFLA algorithm, and the mathematical model of 0-1 KP problem. Then, our proposed CSISFLA for 0-1 KP problems is presented in Section 3. A series of simulation experiments are conducted in Section 4. Some conclusions and comments are made for further research in Section 5.

2. Review of the Related Work

In this section, the model of 0-1 knapsack problem and the basic CS and SFLA are introduced briefly.

2.1. 0-1 Knapsack Problem

The 0-1 knapsack problem, denoted by KP, is a classical optimization problem and it has high theoretical and practical value. Many practical applications can be formulated as a KP, such as cutting stock problems, portfolio optimization, scheduling problems, and cryptography. This problem has been proven to be a NP-hard problem; hence, it cannot be solved in a polynomial time unless P = NP [44]. The 0-1 knapsack problem can be stated as follows: where n is the number of items; w and p represent the weight and profit of item j, respectively. The objective is to select some items so that the total weight does not exceed a given capacity c, while the total profit is maximized. The binary decision variable x , with x = 1 if item i is selected, and x = 0 otherwise is used.

2.2. Cuckoo Search

CS is a relatively new metaheuristic algorithm for solving global optimization problems, which is based on the obligate brood parasitic behavior of some cuckoo species. In addition, this algorithm is enhanced by the so-called Lévy flights rather than by simple isotropic random walks. For simplicity, Yang and Deb used the following three approximate rules [32, 45]: each cuckoo lays only one egg at a time and dumps its egg in a randomly chosen nest; the best nests with high-quality eggs will be carried over to the next generations; the number of available host nests is fixed, and the egg laid by the host bird with a probability p ∈ [0,1]. In this case, the host bird can either throw the egg away or simply abandon the nest and build a completely new nest. The last assumption can be approximated by a fraction p of the n host nests which are replaced by new nests (with new random solutions). New solution X ( is generated as (2) by using a Lévy flight [32]. Lévy flights essentially provide a random walk while their random steps followed a Lévy distribution for large steps which has an infinite variance with an infinite mean. Here the steps essentially form a random walk process with a power-law step-length distribution with a heavy tail as (3): where α > 0 is the step size scaling factor. Generally, we take α = O  (1). The product ⊕ means entry-wise multiplications.

2.3. Shuffled Frog-Leaping Algorithm

The SFLA is a metaheuristic optimization method that imitates the memetic evolution of a group of frogs while casting about for the location that has the maximum amount of available food [46]. SFLA, originally developed by Eusuff and Lansey in 2003, can be applied to handle many complicated optimization problems. In virtue of the beneficial combination of the genetic-based memetic algorithm (MA) and the social behavior-based PSO algorithm, the SFLA has the advantages of global information exchange and local fine search. In SFLA, all virtual frogs are assigned to disjoint subsets of the whole population called memeplex. The different memeplexes are regarded as different cultures of frogs and independently perform local search. The individual frogs in each memeplex have ideas that can be effected by the ideas of other frogs and evolve by means of memetic evolution. After a defined number of memetic evolution steps, ideas are transferred among memeplexes in a shuffling process. The local search and the shuffling processes continue until defined convergence criteria are satisfied [47]. In the SFLA, the initial population P is partitioned into M memeplexes, each containing N frogs (P = M × N). In this process, the ith goes to the jth memeplex where j = i mod M (memeplex numbered from 0). The procedure of evolution of individual frogs contains three frog leapings. The position update is as follows. Firstly, the new position of the frog individual is calculated by If the new position Y is better than the original position X, replace X with Y; else, another new position of this frog will perform in which the global optimal individual B replaces the best individual of kth memeplex B with the following leaping step size: If nonimprovement becomes possible in this case, the new frog is replaced by a randomly generated frog; else replace X with Y: Here, Y is an update of frog's position in one leap. r 1, r 2, and r 3 are random numbers uniformly distributed in [0,1]. B and W are the best and the worst individual of the kth memeplex, respectively. B is the best individual in the whole population. U, L is the maximum and minimum allowed change of frog's position in one leap.

3. Hybrid CS with ISFLA for 0-1 Knapsack Problems

In this section, we will propose a hybrid metaheuristic algorithm integrating cuckoo search and improved shuffled frog-leaping algorithm (CSISFLA) for solving 0-1 knapsack problem. First, the hybrid encoding scheme and repair operator will be introduced. And then improved frog-leaping algorithm along with the framework of proposed CSISFLA will be presented.

3.1. Encoding Scheme

As far as we know, the standard CS algorithm can solve the optimization problems in continuous space. Additionally, the operation of the original CS algorithm is closed to the set of real number, but it does not have the closure property in the binary set {0,1}. Based on above analysis, we utilize hybrid encoding scheme [26] and each cuckoo individual is represented by two tuples 〈x , b 〉 (j = 1,2,…, d), where x works in the auxiliary search space and b performs in the solution space accordingly and d is the dimensionality of solution. Further, Sigmoid function is adopted to transform a real-coded vector X = (x 1,x 2,…,x )T ∈ [−3.0,3.0] to binary vector B = (b 1,b 2,…,b )T ∈ {0,1}. The procedure works as follows: where Sig(x) = 1/(1 + e −) is Sigmoid function. The encoding scheme of the population is depicted in Table 1.
Table 1

Representation of population in CSISFLA.

X 1, B 1X 2, B 2X i, B iX n, B n

3.2. Repair Operator

After evolving a generation, the feasibility of all the generated solutions is taken into consideration. That is, to say, the individuals could be illegal because of violating the constraint conditions. Therefore, a repair procedure is essential to construct illegal individuals. In this paper, an effective greedy transform method (GTM) is introduced to solve this problem [26, 48]. It cannot only effectively repair the infeasible solution but also can optimize the feasible solution. This GTM consists of two phases. The first phase, called repairing phase (RP), checks each solution in order of decreasing p /w and confirms the variable value of one as long as feasibility is not violated. The second phase, called optimizing phase (OP), changes the remaining variable from zero to one until the feasibility is violated. The primary aim of the OP is to transform an abnormal chromosome coding into a normal chromosome, while the RP is to achieve the best chromosome coding.

3.3. Improved Shuffled Frog-Leaping Algorithm

In the evolution of SFLA, new individual is only affected by local optimal individual and the global optimal during the first two frog leapings, respectively. That is to say, there is a lack of information exchange between individuals and memeplexes. In addition, the use of the worst individual is not conducive to quickly obtain the better individuals and quick convergence. When the quality of the solution has not been improved after the first two frog leapings, the SFLA randomly generates a new individual without restriction to replace original individual, which will result in the loss of some valuable information of the superior individual to some extent. Therefore, in order to make up for the defect of the SFLA, an improved shuffled frog-leaping algorithm (ISFLA) is carefully designed and then embedded in the CSISFLA. Compared with SFLA, there are three main improvements. The first slight improvement is that we get rid of sorting of the items according to the fitness value which will decrease in time cost. The second improvement is that we adopt a new frog individual position update formula instead of the first two frog leapings. The idea is inspired by the DE/Best/1/Bin in DE algorithm. Similarly, each frog individual i is represented as a solution X and then the new solution Y is given by where B is the current global best solution found so far. B is the best solution of the kth memeplex. X is an individual of random selection with index of p1 ≠ i and r 2 is random number uniformly distributed in [0,1]. In particular the plus or minus signs are selected with certain probability. The main purpose of improvement in (8) is to quicken convergence rate. The third improvement is to randomly generate new individuals with certain probability instead of unconditional generating new individuals, which takes into consideration the retention of the better individuals in the population. The main step of ISFLA is given in Algorithm 1. In Algorithm 1, P is the size of the population. M is the number of memeplex. D is the dimension of decision variables. And r 1 is a random real number uniformly distributed in (0, 1). And r 2, r 3, r 4, and p are all D-dimensional random vectors and each dimension is uniformly distributed in (0, 1). In particular, p is called probability of mutation which controls the probability of individual random initialization.
Algorithm 1

Improved shuffled frog-leaping algorithm.

3.4. The Frame of CSISFLA

In this section, we will demonstrate how we combine the well-designed ISFLA with Lévy flights to form an effective CSISFLA. The proposed algorithm does not change the main search mechanism of CS and SFLA. In the iterative process of the whole population, Lévy flights are firstly performed and then frog-leaping operator is adopted in each memeplex. Therefore, the strong exploration abilities in global area of the original CS and the exploitation abilities in local region of ISFLA can be fully developed. The CSISFLA architecture is explained in Figure 1.
Figure 1

The architecture of CSISFLA algorithm.

3.5. CSISFLA Algorithm for 0-1 Knapsack Problems

Through the design above carefully, the pseudocode of CSISFLA for 0-1 knapsack problems is described as follows (see Algorithm 2). It can be analyzed that there are essentially three main processes besides the initialization process. Firstly, Lévy flights are executed to get a cuckoo randomly or generate a solution. The random walk via Lévy flights is much more efficient in exploring the search space owing to its longer step length. In addition, some of the new solutions are generated by Lévy flights around the best solution, which can speed up the local search. Then ISFLA is performed in order to exploit the local area efficiently. Here, we regard the frog-leaping process as the process of cuckoo laying egg in a nest. The new nest generated with a probability p is far enough from the current best solution, which enables CSISFLA to avoid being trapped into local optimum. Finally, when an infeasible solution is generated, a repair procedure is adopted to keep feasibility and, moreover, optimize the feasible solution. Since the algorithm can well balance the exploitation and exploration, it expects to obtain solutions with satisfactory quality.
Algorithm 2

The main procedure of CSISFLA algorithm.

3.6. Algorithm Complexity

CSISFLA is composed of three stages: the sorting by value-to-weight ratio, the initialization, and the iterative search. The quick sorting has time complexity O (Plog⁡ (P)). The generation of the initial cuckoo nests has time complexity O (P × D). The iterative search consists of four steps (comment statements in Algorithm 2), and so forth, the Lévy flight, the first frog leaping, generate new individual and random selection which costs the same time O (D). In summary, the overall complexity of the proposed CSISFLA is O (Plog⁡ (P)) + O (P × D) + O (D) = O (Plog⁡ (P)) + O (P × D). It does not change compared with the original CS algorithm.

4. Simulation Experiments

4.1. Experimental Data Set

In existent researching files, cases studies and research of knapsack problems are about small-scale to moderate-scale problems. However, in real-world applications, problems are typically large-scale with thousands or even millions of design variables. In addition, the complexity of KP problem is greatly affected by the correlation between profits and weights [49-51]. However, few scholars pay close attention to the correlation between the weight and the value of the items. To test the validity of the algorithm for different types of instances, we adopt uncorrelated, weakly correlated, strongly correlated, multiple strongly correlated, profit ceiling, and circle data sets with different dimension. The problems are described as follows: uncorrelated instances: the weights w and the profits p are random integers uniformly distributed in [10,100]; weakly correlated instances: the weights w are random integers uniformly distributed in [10,100], and the profits p are random integer uniformly distributed in [w − 10, w + 10]; strongly correlated instances: the weights w are random integers uniformly distributed in [10,100] and the profits p are set to w + 10; multiple strongly correlated instances: the weights w are randomly distributed in [10,100]. If the weight w is divisible by 6, then we set the p = w + 30 otherwise set it to p = w + 20; profit ceiling instances: the weights w are randomly distributed in [10,100] and the profits p are set to p = 3⌈w /3⌉; circle instances: the weights w are randomly distributed in [10,100] and the profits p are set to . Choosing d = 2/3, R = 10. For each data set, we set the value of the capacity. Consider c = 0.75∑ w . Figures 2, 3, 4, 5, 6, and 7 illustrate six types of instances of 200 items, respectively.
Figure 2

Uncorrelated items.

Figure 3

Weakly correlated items.

Figure 4

Strongly correlated items.

Figure 5

Multiple strongly correlated items.

Figure 6

Profit ceiling items.

Figure 7

Circle items.

The KP instances in this study are described in Table 2.
Table 2

Knapsack problem instances.

ProblemCorrelationDimensionTarget weightTotal weightTotal values
KP1 Uncorrelated150647186288111
KP2 Uncorrelated20083281110410865
KP3 Uncorrelated300123831651116630
KP4 Uncorrelated500203632715028705
KP5 Uncorrelated800333674448944005
KP6 Uncorrelated1000419485593054764
KP7 Uncorrelated1200494856598066816
KP8 Weakly correlated150640385388504
KP9 Weakly correlated20083581114411051
KP10 Weakly correlated300125541673916778
KP11 Weakly correlated500207582767727821
KP12 Weakly correlated800333674448944491
KP13 Weakly correlated1000418495579955683
KP14 Weakly correlated1200498086641156811
KP15 Strongly correlated300122471632919329
KP16 Strongly correlated500213052840733406
KP17 Strongly correlated800333674448952489
KP18 Strongly correlated1000408835451164510
KP19 Strongly correlated1200504306724079240
KP20 Multiple strongly correlated300129081721123651
KP21 Multiple strongly correlated500202592701237903
KP22 Multiple strongly correlated800327674368961140
KP23 Multiple strongly correlated1000424425658977940
KP24 Multiple strongly correlated1200502226696392653
KP25 Profit ceiling300126661688817181
KP26 Profit ceiling500198112641526913
KP27 Profit ceiling800320114268143497
KP28 Profit ceiling1000422535633757381
KP29 Profit ceiling1200502086694468157
KP30 Circle300125541673926448
KP31 Circle500208122774943880
KP32 Circle800325814344169527
KP33 Circle1000421075614388220
KP34 Circle12004922065627104287

4.2. The Selection on the Value of M and N

The CSISFLA has some control parameters that affect its performance. In our experiments, we investigate thoroughly the number of subgroups M and the number of individuals in each subgroup N. The below three test instances are used to study the effect of M and N on the performance of the proposed algorithm. Firstly, M is set to 2, and then three levels of 10, 15, and 20 are considered for N (accordingly, the size of population is 2 × 10, 2 × 15, and 2 × 20). Secondly, a fixed individual number of each subgroup is 10, and the value of M is 2, 3, and 4, respectively. Results are summarized in Table 3.
Table 3

The effect of M and N on the performance of the CSISFLA.

Instance N M = 2 M N = 10
BestWorstMeanSTDBestWorstMeanSTD
KP9 108727 8704 87115.52 8727 870487115.5
15 8728 8701 8715 6.8 38725870187137.0
20 8730 8702 8718 6.548726 8708 8717 6.3

KP10 101315213124131408.721315213124131408.7
15 13168 131201314412.6313167 13131 13146 8.2
20 13174 13126 13148 13.34 13168 13128 13148 9.4

KP11 1021820217372177322.1221820217372177322.1
1521827 21756 21786 17.3 3 21840 217352178324.6
2021814 21757 21778 15.4 4 21848 21742 21788 23.5
As expected, with the increase of the individual number in the population, it is an inevitable consequence that there are more opportunities to obtain the optimal solution. This issue can be indicated by bold data in Table 3. In order to get a reasonable quality under the condition of inexpensive computational costs, we use N = 10 and M = 4 in the rest experiments.

4.3. The Selection on the Value of p

In this subsection, the effect of p on the performance of the CSISFLA is carefully investigated. We select two uncorrelated instances (KP1, KP2) and two weakly correlated instances (KP8, KP9) as the test instances for parameter setting experiment of p . For each instance, every test is run 30 times. We use N = 10, M = 4, and the maximum time of iterations is set to 5 seconds. Table 4 gives the optimization results of the CSISFLA using different values for p .
Table 4

The effect of p on the performance of the CSISFLA.

Instance00.050.10.150.20.30.40.50.60.70.80.91.0
KP1
 Best7474 7475 7475 7475 7475 7474 7475 747474747474747374747459
 Worst743074697468 7471 7471 74637457745174517446743774277407
 Mean74617473 7474 7474 747374717470746874687461745574487436
 STD12.601.501.57 0.93 1.273.574.966.035.878.8310.1111.1713.88
KP2
 Best 9865 9865 9865 9865 986398649860985998509847984498439842
 Worst9821 9847 98459844983998239830981898049778977597689757
 Mean9847 9858 98569857985298489847984198339830981298109783
 STD11.965.756.12 5.32 6.8410.607.9911.8912.3516.8621.9221.1220.24
KP8
 Best66766674667366726671667266726671 6678 6666666666626654
 Worst665866626663 6665 666266636662665766556650665266456642
 Mean6668 6671 66696669666866686668666466646659665866526647
 STD4.592.952.59 2.04 2.442.792.394.174.454.063.884.273.17
KP9
 Best8730 8734 8734 8728873187208723871687128710870787018688
 Worst 8707 870387058701870087028695868486828675867786648655
 Mean8716 8718 8718 8715871487118707870286978693869086828676
 STD6.238.796.666.857.45 4.59 7.207.977.509.757.2710.067.76
From the results of Table 4, it is not difficult to observe that the probability of mutation with 0.05 ≤ p ≤ 0.4 is more suitable for all test instances which can be seen from data in bold in Table 3. In addition, the optimal solution dwindles steadily with the change of p from 0.5 to 1.0 and the worst results of four evaluation criteria are obtained when p = 1. Similarly, the performance of the CSISFLA is also poor when p is 0. As we have expected, 0 means that the position update in memeplex is completed entirely by the first Leapfrog, which cannot effectively ensure the diversity of the entire population, leading to the CSISFLA more easily fall into the local optimum, and 1 means that new individuals randomly generated without any restrictions which results in slow convergence. Generally speaking, using a small value of p is beneficial to strengthen the convergence ability and stability of the CSISFLA. The performance of the algorithm is the best when p = 0.15, so we will set p = 0.15 for the following experiments.

4.4. Experimental Setup and Parameters Setting

In this paper, in order to test the optimization ability of CSISFLA and further investigate effectiveness of the algorithms for different types of instance, we adopt a set of 34 knapsack problems (KP1–KP34). We compared the performance of CSISFLA with (a) GA, (b) DE, and (c) classical CS. In the experiments, the parameters setting are shown in Table 5.
Table 5

Parameter settings of GA, DE, CS, and CSISFLA on 0-1 knapsack problems.

AlgorithmParameterValue
GA [41] Population size 100
Crossover probability 0.6
Mutation probability 0.001
DE [42, 43] Population size 100
Crossover probability 0.9
Amplification factor 0.3
CS [24] Population size 40
p a 0.25
CSISFLA M 4
N 10
p m 0.15
In order to make a fair comparison, all computational experiments are conducted with Visual C++ 6.0. The test environment is set up on a PC with AMD Athlon(tm) II X2 250 Processor 3.01 GHz, 1.75 G RAM, running on Windows XP. The experiment on each instance was repeated 30 times independently. Further, best solution, worst solution, mean, median, and standard deviation (STD) for all the solutions are given in related tables. In addition, the maximum run-time was set to 5 seconds for the instances with dimension less than 500, and it was set to 8 seconds for other instances.

4.5. The Experimental Results and Analysis

We do experiment on 7 uncorrelated instances, 7 weakly correlated instances, and 5 other types of instances, respectively. The numerical results are given in Tables 6–11. The best values are emphasized in boldface. In addition, comparisons of the best profits obtained from the CSISFLA with those obtained from GA, DE, and CS for six KP instances with 1200 items are shown in Figures 8, 9, 10, 11, 12, and 13. Specifically, the convergence curves of four algorithms on six KP instances with 1200 items are also drawn in Figures 14, 15, 16, 17, 18, and 19. Through our careful observation, it can be analyzed as follows.
Table 6

Experimental results of four algorithms with uncorrelated KP instances.

InstanceAlgorithmBestWorstMeanMedianSTD
KP1 GA731669787200720875.78
DE 7475 7433747174737.68
CS747273587403740527.82
CSISFLA 7475 7467 7473 7474 1.56

KP2 GA967392279503950797.39
DE 9865 97519854 9865 22.52
CS984896789737973433.22
CSISFLA 9865 9837 9856 9858 7.23

KP3 GA15022142751475614795158.91
DE 15334 15088152871530154.45
CS1522415024150921508151.37
CSISFLA15327 15248 15297 15302 18.48

KP4 GA25882252122549825493150.68
DE26333257512609926096135.88
CS26208257862593625911103.4
CSISFLA 26360 26193 26284 26277 38.54

KP5 GA39528384623897639014243.62
DE39652392153941039399113.28
CS40223394163956539514179.98
CSISFLA 40290 39885 40072 40081 91.97

KP6 GA49072478354848348570316.62
DE49246488354898948979101.11
CS49767490244916449142143.08
CSISFLA 49893 49567 49744 49737 97.52

KP7 GA59793583515913559225370.86
DE59932594885970759727 110.39
CS60629597085993959884166.43
CSISFLA 60779 60264 60443 60420 130.56
Table 11

Experimental results of four algorithms with circle KP instances.

InstanceAlgorithmBestWorstMeanMedianSTD
KP30 GA2119420899210862109671.44
DE2133321192212642127732.46
CS21333211942126121261 18.57
CSISFLA 21333 21263 21300 21295 34.04

KP31 GA3526234982351123512482.25
DE3534335184352473526738.08
CS3534535271352973527731.29
CSISFLA 35414 35342 35354 35345 23.23

KP32 GA55976554515574655771116.83
DE5606355914559645595444.95
CS 56280 55988560575606155.01
CSISFLA56273 56130 56185 56201 38.65

KP33 GA70739702477048770456113.53
DE70806706417069670684 38.21
CS7091570729707897079742.50
CSISFLA 71008 70867 70924 70939 41.17

KP34 GA83969833398372383757142.75
DE8404083820839128389956.64
CS 84645 839548405584033121.94
CSISFLA84244 84099 84175 84181 38.36
Figure 8

The best profits obtained in 30 runs for KP7.

Figure 9

The best profits obtained in 30 runs for KP14.

Figure 10

The best profits obtained in 30 runs for KP19.

Figure 11

The best profits obtained in 30 runs for KP24.

Figure 12

The best profits obtained in 30 runs for KP29.

Figure 13

The best profits obtained in 30 runs for KP34.

Figure 14

The convergence graphs of KP7.

Figure 15

The convergence graphs of KP14.

Figure 16

The convergence graphs of KP19.

Figure 17

The convergence graphs of KP24.

Figure 18

The convergence graphs of KP29.

Figure 19

The convergence graphs of KP34.

Table 6 shows that CSISFLA outperforms GA, DE, and CS on almost all the uncorrelated knapsack instances in terms of computation accuracy and robustness. In particular, the best solution found by CSISFLA is slightly inferior to that obtained by DE on KP3. On closer inspection, “STD” is much smaller than that of the other algorithms except for KP7, which indicates the good stability of the CSISFLA and superior approximation ability. From Table 7, it can be seen that DE obtained the best, mean, and median results for the first four cases, and CS attained the best results for the last three cases. Although the optimal solutions obtained by the CSISFLA are worse than DE or CS, the CSISFLA obtained the worst, median, and STD results in KP12–KP14, which still can indicate that the CSISFLA has better stability. Above all, the well-known NFL theorem [52] has stated clearly that there is no heuristic algorithm best suited for solving all optimization problems. Unfortunately, although weakly correlated knapsack problems are closer to the real world situations [49], the CSISFLA does not appear clearly superior to the other two algorithms in solving such knapsack problems.
Table 7

Experimental results of four algorithms with weakly correlated KP instances.

InstanceAlgorithmBestWorstMeanMedianSTD
KP8 GA662765316593659320.63
DE 6676 6657 6674 6676 4.80
CS66606637664866466.79
CSISFLA6673 6663 66686668 2.23

KP9 GA865885018588859033.38
DE 8743 8743 8743 8743 0.00
CS871786448676867118.23
CSISFLA87288701871487146.87

KP10 GA1306212939129971299130.64
DE 13202 13158 13186 13186 9.76
CS1315713069130941308721.91
CSISFLA1316813120131451314511.90

KP11 GA2167121470215712157648.85
DE 21951 21745 21858 21859 37.61
CS2193521670217462172276.53
CSISFLA21827 21756 2178821787 16.66

KP12 GA3458734314344883449963.23
DE3481434578347213471864.50
CS 34987 346213469734654100.38
CSISFLA34818 34721 34760 34758 22.87

KP13 GA4324142938430824307375.51
DE4332743162432174321143.64
CS 43737 432164334043264166.53
CSISFLA43409 43312 43367 43368 27.23

KP14 GA51472504145105851135265.56
DE51947514445160051569108.83
CS 53333 516015183151788299.35
CSISFLA52403 52077 52267 52264 86.19
Obviously, in point of search accuracy and convergence speed, it can be seen from Table 8 that CSISFLA outperforms GA, DE, and CS on all five strongly correlated knapsack problems. If anything, the STD values tell us that CSISFLA is only inferior to CS.
Table 8

Experimental results of four algorithms with strongly correlated KP instances.

InstanceAlgorithmBestWorstMeanMedianSTD
KP15 GA1478514692147541476225.93
DE147971478114789147874.90
CS148041479114797 14797 2.43
CSISFLA 14807 14795 14798 14797 3.46

KP16 GA2548625402254582546521.61
DE255022548125492254934.21
CS25514255022550625505 3.49
CSISFLA 25515 25505 25510 25512 3.94

KP17 GA4008739975400394004128.33
DE401114006840089400888.66
CS40107400964010340105 3.88
CSISFLA 40117 40098 40111 40113 5.12

KP18 GA4933249225493004930927.26
DE493634933349346493457.50
CS49380493504936449363 7.04
CSISFLA 49393 49362 49373 49373 7.90

KP19 GA6052060418604826048926.62
DE605406050160519605198.55
CS605586053060542605406.77
CSISFLA 60562 60539 60549 60550 5.70
Similar results were found from Tables 9, 10, and 11 and it can be inferred that CSISFLA can easily yield superior results compared with GA, DE, and CS. The series of experimental results confirm convincingly the superiority and effectiveness of CSISFLA.
Table 9

Experimental results of four algorithms with multiple strongly correlated KP instances.

InstanceAlgorithmBestWorstMeanMedianSTD
KP20 GA1834618172182841828838.39
DE1838718335183541834815.25
CS18386183551836818368 4.73
CSISFLA 18388 18368 18381 18386 8.03

KP21 GA2952529387294612946231.97
DE2954829488295192952014.10
CS2958929527295552954913.94
CSISFLA 29609 29562 29581 29585 12.38

KP22 GA4764547494475684757539.72
DE4770447620476594765720.68
CS4772747673476964769515.09
CSISFLA 47757 47697 47732 47736 13.02

KP23 GA6052960312604556046347.39
DE60572605086053460530 13.98
CS6060760540605766057416.96
CSISFLA 60650 60579 60615 60612 15.75

KP24 GA7206371725719147191764.42
DE7207271973720187201819.38
CS72094720317205872057 15.93
CSISFLA 72151 72070 72112 72111 21.20
Table 10

Experimental results of four algorithms with profit ceiling KP instances.

InstanceAlgorithmBestWorstMeanMedianSTD
KP25 GA129571294812955129572.53
DE129571295112953129541.83
CS129571295412957129570.76
CSISFLA 12957 12957 12957 12957 0.00

KP26 GA202952026820285202867.37
DE203012029220294202942.17
CS20304202952029920298 1.86
CSISFLA 20307 20298 20304 20304 2.28

KP27 GA327963276932785327876.99
DE32802327933279732796 2.63
CS328113279932803328023.12
CSISFLA 32820 32808 32812 32811 3.34

KP28 GA432484321543234432368.76
DE432574324543249432483.57
CS432694325143257432544.41
CSISFLA 43272 43260 43266 43266 2.88

KP29 GA513785134851364513667.25
DE51384513725137851378 3.04
CS 51399 5137851385513844.32
CSISFLA 51399 51390 51396 51396 3.10
Figures 8–13 show a comparison of the best profits obtained by the four algorithms for six types of 1200 items. Figures 14–19 illustrate the average convergence curves of all the algorithms in 30 runs where we can observe that CS and CSISFLA usually show the almost same starting point. However, CSISFLA surpasses CS in point of the accuracy and convergence speed. CS performs the second best in hitting the optimum. DE shows premature phenomenon in the evolution and does not offer satisfactory performance along with the extending of the problem. Based on previous analyses, we can draw a conclusion that the superiority of CSISFLA over GA, DE, and CS in solving six types of KP instances is quite indubitable. In general, CS is slightly inferior to CSISFLA, so the next best is CS. DE and GA perform the third-best and the fourth-best, respectively.

5. Conclusions

In this paper, we proposed a novel hybrid cuckoo search algorithm with improved shuffled frog-leaping algorithm, called CSISFLA, for solving 0-1 knapsack problems. Compared with the basic CS algorithm, the improvement of CSISFLA has several advantages. First, we specially designed an improved frog-leap operator, which not only retains the effect of the global optimal information on the frog leaping but also strengthens information exchange between frog individuals. Additionally, new individuals randomly generated with mutation rate. Second, we presented a novel CS model which is in an excellent combination with the rapid exploration of the global search space by Lévy flight and the fine exploitation of the local region by frog-leap operator. Third, CSISFLA employs hybrid encoding scheme; that is, to say, it conducts active searches in continuous real space, while the consequences are used to constitute the new solution in the binary space. Fourth, CSISFLA uses an effective GTM to assure the feasibility of solutions. The computational results show that CSISFLA outperforms the GA, DE, and CS in solution quality. Further, compared with ICS [26], the CSISFLA can be regarded as a combination of several algorithms and secondly the KP instances are more complex. The future work is to design more effective CS method for solving complex 0-1 KP and to apply the hybrid CS for solving other kinds of combinatorial optimization problems, multidimensional knapsack problem (MKP), and traveling salesman problem (TSP).
  3 in total

1.  A hybrid metaheuristic DE/CS algorithm for UCAV three-dimension path planning.

Authors:  Gaige Wang; Lihong Guo; Hong Duan; Heqi Wang; Luo Liu; Mingzhen Shao
Journal:  ScientificWorldJournal       Date:  2012-10-21

2.  An improved hybrid encoding cuckoo search algorithm for 0-1 knapsack problems.

Authors:  Yanhong Feng; Ke Jia; Yichao He
Journal:  Comput Intell Neurosci       Date:  2014-01-12

3.  An effective hybrid firefly algorithm with harmony search for global numerical optimization.

Authors:  Lihong Guo; Gai-Ge Wang; Heqi Wang; Dinan Wang
Journal:  ScientificWorldJournal       Date:  2013-11-20
  3 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.