Dynamic Inertia Weight Binary Bat Algorithm with Neighborhood Search
Dynamic Inertia Weight Binary Bat Algorithm with Neighborhood Search
Xingwang Huang,1,2 Xuewen Zeng,1 and Rui Han1
1National Network New Media Engineering Research Center, Institute of Acoustics, Chinese Academy of Sciences, Beijing 100190, China
2University of Chinese Academy of Sciences, Beijing 100049, China
Correspondence should be addressed to Rui Han; nc.ca.psd@rnah
Received 27 February 2017; Accepted 20 April 2017; Published 28 May 2017
Academic Editor: Michael Schmuker
Copyright © 2017 Xingwang Huang et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
Binary bat algorithm (BBA) is a binary version of the bat algorithm (BA). It has been proven that BBA is competitive compared to other binary heuristic algorithms. Since the update processes of velocity in the algorithm are consistent with BA, in some cases, this algorithm also faces the premature convergence problem. This paper proposes an improved binary bat algorithm (IBBA) to solve this problem. To evaluate the performance of IBBA, standard benchmark functions and zero-one knapsack problems have been employed. The numeric results obtained by benchmark functions experiment prove that the proposed approach greatly outperforms the original BBA and binary particle swarm optimization (BPSO). Compared with several other heuristic algorithms on zero-one knapsack problems, it also verifies that the proposed algorithm is more able to avoid local minima.
There are many optimization problems with binary search space. And many of them are high dimensional. Thus, it is infeasible to solve them with exhaustive method. So as to optimize these problems, such as unit commitment , feature selection [2, 3], task scheduling [4, 5], and 0-1 knapsack problem [6, 7], binary algorithms are proposed to generate binary solutions. For example, Mirjalili et al.  adapted the standard continuous BA algorithm to be applied to binary spaces and then BBA combined with -Nearest Neighbor (KNN, ) method was used to solve feature selection problem . BBA can provide competitive performance but, in some cases, it may get stuck to local minima. To solve this issue, an improved binary bat algorithm, named IBBA, is proposed. IBBA will carry out a more diversified search process. The main contributions of the paper can be summarized as follows:(1)An improved high-performance binary bat algorithm is proposed for binary problems. Using the neighbor bat and dynamic inertia weight strategy, the proposed approach can be more able to avoid being trapped into local minima.(2)To evaluate its performance, the proposed IBBA and several other algorithms are implemented on benchmark functions and zero-one knapsack problems. The results obtained prove that IBBA outperforms the other algorithms.
The organization of the paper is as follows: a compact overview of BA and BBA is given in Section 2. A literature review on inertia weight strategies is also provided in this section. Section 3 presents the improved structure of IBBA. The experimental results of benchmark functions and zero-one knapsack problems are demonstrated in Sections 4 and 5, respectively. And the reason that the performance of IBBA is better than other algorithms is given in Section 6. Finally, conclusion is made in Section 7.
This section provides a brief overview of BA and BBA. A literature review on inertia weight strategies is also presented.
2.1. The Bat Algorithm
Inspired by the echolocation behavior of bats, Yang proposed the bat algorithm . When bats chase preys, they will decrease the loudness and increase the frequency of emitted ultrasonic sound. These characteristics of real bats have been used in developing the BA. These basic steps of BA have been mathematically described as follows .
In the BA, each bat has three vectors, including a frequency vector, a velocity vector, and a position vector that are updated at time step as (1), (2), and (3):where represents the best position obtained so far and represents the frequency of th bat which is updated as follows:where in the range of is a random vector drawn from a uniform distribution. From (1) and (3), it is obvious that different frequencies promote the exploration capability of bats to the optimal solution.
These equations, to a certain extent, can guarantee the exploitation capability of the BA. However, to perform the intensification better, a random walk operation has also been employed as follows:where means one solution selected randomly among the current best solutions, is a randomly selected number in the range of , and indicates the average loudness of all bats at this time step. The pseudocode of BA algorithm is demonstrated in Algorithm 1. Note that rand is a random number uniformly distributed in the range . To an extent, BA can be considered as a balanced combination of global and intensive local search. The pulse emission rate () and loudness () control the balancing between these two search techniques. As increases, artificial bats tend to perform a diversification rather than intensification. These two parameters mentioned above are updated as follows:where and are constants and has the same meaning of the cooling factor in SA . To guarantee that the artificial bats are moving toward the optimal solutions, both loudness and emission rate are updated when the better solutions are found. For any ,For simplicity, can be used.
Algorithm 1: Pseudocode of bat algorithm.
2.2. Binary Bat Algorithm
The binary bat algorithm (BBA) was proposed by Mirjalili et al.  to solve optimization problems with binary search space. The structure of BBA is almost the same as the original BA in which the velocity and frequency are defined in continuous space. BBA makes two changes to the original BA:(i)The vector of position is no longer a continuous-valued vector but a bit string.(ii)The random operation demonstrated by (4) is no longer suitable to binary search space. Instead, a simpler operation is adopted.
The position update equation for BBA changes towhereand and indicate the position and velocity of th artificial bat at iteration in th dimension and represents the complement of .
The operation demonstrated by (4) for BBA changes towhere still denotes a solution selected randomly from the current best solutions.
2.3. Inertia Weight Strategies
In some heuristic algorithms, especially PSO , inertia weight strategy plays an important role in the process of keeping balance between global search and local search process. The inertia weight strategy determines the contribution proportion of a particle’s old velocity to its new velocity at the current time step. Shi and Eberhart  proposed the concept of inertia weight by adopting constant inertia weight strategy and demonstrated that a large inertia weight enhances the exploration while a small inertia weight enhances the exploitation. Further, various dynamic inertia weight strategies have been proposed which can improve the capabilities of PSO and they can be categorized into three classes: constant and random inertia weight strategies, time-varying inertia weight strategies, and adaptive inertia weight strategies. A compact literature review of inertia weight strategies is presented in subsequent paragraphs.
Eberhart and Shi  presented a random inertia weight strategy which was demonstrated to be more suitable for dynamic problems. Khan et al.  proposed a modified PSO by introducing a mutation mechanism and using dynamic algorithm parameters. To increase the diversity of the particles, the inertia weight of each particle adopts a random updating formula.
Most of the PSO variants employed time-varying inertia weight strategies in which the value of the inertia weight is adjusted based on the iteration number. In , a linear decreasing variant of inertia weight was proposed and was illustrated to be effective in enhancing the fine tuning performance of the PSO. Inspired by the idea of decreasing the inertia weight over time step, a nonlinear decreasing inertia weight strategy was proposed . Gao et al.  presented a novel PSO variant which combined chaos mutation operator with the logarithm decreasing inertia weight. Based on the idea of decreasing inertia weight, Chen et al.  proposed two natural exponent inertia weight strategies to solve the continuous optimization problems.
Adaptive inertia weight strategies is another research trend of inertia weight strategies which monitor the search situation and adjust the inertia weight value according to one or more feedback parameters. Nickabadi et al.  proposed a new adaptive inertia weight approach that employs the success rate of the swarm as the feedback parameter to ascertain the particle’s situation in the search space. Zhan et al.  presented an adaptive particle swarm optimization (APSO) which provides better search efficiency than classical PSO. Yang et al.  used speed factor and aggregation degree factor to adapt the value of inertia weight.
3. Improved Binary Bat Algorithm
BA is an algorithm combined with many merits of previous proposed heuristic algorithms, such as PSO  and SA . Therefore, it stands for the reason that the update process of velocity and location in the BA has many similarities with PSO. When the velocity update equation (1) is analyzed, it is obvious that this equation consists of two parts. The first item () denotes the velocity of population and the second item () controls the velocity of the th position () with guidance of the global best solution (). The first and second items of the equation affect the algorithm so that it performs global and local search, respectively. It has been proven that the first item of (1) may reduce the convergence rate rapidly and the second item of (1) may result in premature convergence problem . To solve this problem, some improved BA algorithms have been proposed recently [22–25]. Due to the fact that the structure of BBA is effectively the same as the original BA, BBA is not good enough at exploration and exploitation, too.
EBA  illustrates that the algorithm could produce better solutions with the guidance of the neighbor bat (th solution). For this purpose, inspired by , the velocity update equation of original BBA is modified as follows:where denotes the inertia weight factor which balances local and global search intensity of the th solution by controlling the value of old velocity , represents one of the best solutions randomly selected from the population (), is self-adaptive learning factor of global best solution () ranging from 0 to 1, and therefore, , which is a learning factor of th solution, ranges from 1 to 0. Since the th solution information is used to guide the th solution, the algorithm can be more able to avoid local minima. As is increased, the effect of the global best solution () becomes higher than the th neighbor solution () and vice versa. The update equation for is shown as follows:where denotes initial impact factor of , represents the maximum number of iterations, iter indicates the current number of iterations, and indicates a nonlinear modulation index. As iter is increased, will increase from to 1 nonlinearly and will decrease from () to 0 correspondingly. With a small and a large , bats are allowed to fly around the search space, instead of flying toward the swarm best. On the other hand, a large and a small allow the bats to converge to the global optimum solution in the latter stages of the search process. Therefore, the proposed approach can effectively control the global search and enhance convergence to the global best solution during the latter part of the optimization. Thus, the solution can switch from exploration to exploitation. The states of and have been illustrated in Figure 1 when and are 0.6 and 2, respectively.
Figure 1: The changes of and with iterations.
Inspired by , dynamic inertia weight strategy is used to control the magnitude of the velocity. This strategy is illustrated as follows:where indicates the total number of iterations, iter denotes the current number of iterations, maximal inertia values are represented by , and is a constant larger than 1.
It is crucial to balance the large-scale exploration search and exploitation search. Once the algorithm locates the approximate position of global optima, refined local exploitation search capabilities need to be enhanced to get global optimum. To dynamically control the transformation point of global search and local search, an adaptive strategy is designed. If the current global best solution is not improved after iterations, the algorithm switches to intensive exploitation search with the smaller or continues exploration search with the current . The strategy is defined as follows:where and represent the th and th taken values of , respectively, and indicates an interval of definite iterations.
According to the descriptions mentioned above, the benefit of the proposed improved binary bat algorithm (IBBA) is that it contributes to the dispersion of the solutions into binary search space. In addition, more accurate results can be obtained.
4. Benchmark Experiment
To verify the performance of proposed algorithm, thirteen benchmark functions  are employed. These test functions contain two different groups: unimodal and multimodal functions. Tables 1 and 2 demonstrate the benchmark functions used, respectively, where range means the search boundary of the function. The dimensions of all test functions are 5. The global minimum values of all benchmark functions used except are 0 while the global minimum value of is ().
Table 1: Unimodal benchmark functions.
Table 2: Multimodal benchmark functions.
For this simulation, 15 bits are used to represent each continuous variable in binary. Thus, the dimension of generating bit vector for each benchmark function is calculated as follows:where is the dimension of each artificial bat in IBBA and represents the dimension of a particular benchmark function. Therefore, the dimensions of the corresponding binary problems are ().
The simulation focused on comparing the performance of the IBBA, BBA , and BPSO . Some basic parameters should be initialized before running these algorithms. The initial parameters for these algorithms are demonstrated in Table 3. Tables 4 and 5 present the experimental results. The obtained results of each algorithm are averaged over 30 independent runs, and best results are denoted in bold type. In addition, the mean (Mean), standard deviation value (SD), and medium value (Med) of the optimal solution in the last iteration are presented.
Table 3: Initial parameters for IBBA, BBA, and BPSO.
Table 4: Performance comparison on unimodal benchmark functions.
Table 5: Performance comparison on multimodal benchmark functions.
To judge whether IBBA is significantly different from BBA and BPSO or not, we carried out the statistical Wilcoxon’s rank-sum test  at a 5% significance level. The values calculated in Wilcoxon’s rank-sum test comparing IBBA and other algorithms over all the benchmark functions are summarized in Tables 6 and 7. In the demonstrated tables, NA means “Not Applicable” which indicates that the corresponding approach could not statistically compare with itself in the Wilcoxon’s rank-sum test. According to Derrac et al. , if a value is < 0.05, it can be considered as strong evidence against the null hypothesis. In other words, it means that the compared algorithms have significant difference from each other.
Table 6: values on unimodal benchmark functions.
Table 7: values on multimodal benchmark functions.
4.1. Unimodal Benchmark Functions
Unimodal benchmark functions are effective to quantify the convergence speed. Tables 4 and 6 summarize the results for these benchmark functions. It can be observed from Table 4 that IBBA could get the most optimal values. values in Table 6 also point out that IBBA outperforms the other algorithms significantly. The results obtained prove that the proposed IBBA has excellent performance in finding the global optimal solution of unimodal functions.
Figures 2–8 illustrate the averaged convergence curves of algorithms on unimodal benchmark functions. As these curves show, the accuracy of the IBBA is much better than other algorithms on these unimodal benchmark functions in terms of the mean, standard deviation value, and median value of the results. According to Tables 4 and 6 and Figures 2–8, it can be concluded that the IBBA is capable of finding the optimal solution with considerable fast convergence rate in unimodal benchmark functions.
Figure 2: The averaged convergence curve of .
Figure 3: The averaged convergence curve of .
Figure 4: The averaged convergence curve of .
Figure 5: The averaged convergence curve of .
Figure 6: The averaged convergence curve of .
Figure 7: The averaged convergence curve of .
Figure 8: The averaged convergence curve of .
4.2. Multimodal Benchmark Functions
Multimodal benchmark functions are usually used to detect whether the algorithm faces premature convergence problem. Tables 5 and 7 provide the results of the multimodal benchmark functions. Table 5 shows that IBBA could get the best results in all cases. And values in Table 7 indicate that, compared to other algorithms, IBBA achieves significant performance boost in all the multimodal benchmark functions. Therefore, this is evidence that IBBA has the best capability of avoiding to be trapped in local minima.
The averaged convergence curves of all algorithms when they deal with multimodal benchmark functions are demonstrated in Figures 9–14. It can be observed from these figures that IBBA also has a significant improvement on the accuracy than other algorithms on multimodal benchmark functions. According to Tables 5 and 7 and Figures 9–14, results obtained prove that the IBBA is able to avoid local minima.
Figure 9: The averaged convergence curve of .
Figure 10: The averaged convergence curve of .
Figure 11: The averaged convergence curve of .
Figure 12: The averaged convergence curve of .
Figure 13: The averaged convergence curve of .
Figure 14: The averaged convergence curve of .
According to the comparison and discussion mentioned above, it can be concluded that the proposed algorithm has huge merit among the binary algorithms. The next section presents the performance of IBBA in solving the zero-one knapsack problems.
5. Zero-One Knapsack Problem
Zero-one knapsack problem is a classical combinatorial optimization problem  and it has plenty of immediate applications in industry and financial management. It is assumed that, in this problem, there are many items waiting to be carried in knapsack and each item is with a specific benefit and weight. The aim of this problem is to maximize the total profit in the knapsack while the total weight in the knapsack is not more than the given limit. The zero-one knapsack problem can be modeled as follows :where is the number of items and to have a benefit and weight . The maximum weight that this knapsack can carry is . Assume that all benefits and weights are nonnegative. In this study, five typical knapsack problem datasets are used . These test datasets are listed in Table 8, where Dim represents the dimension of knapsack problems, parameters present the information of weight, price, and weight limit, and Opt indicates the optimal value of corresponding knapsack problem. Table 3 is used to set the initial parameters of each algorithm, too. The averaged results obtained are listed in Table 9 and optimal results are denoted in bold type.
Table 8: Five typical test datasets of zero-one knapsack problem.
Table 9: Performance comparison on zero-one knapsack problems over 30 independent runs.
According to Table 9, it can be observed that, in all cases, IBBA performs better than the other two algorithms. The maximum results of IBBA in four out of five test datasets are equal to the optimal values of corresponding knapsack problems. In the remaining test datasets (), the maximum result of IBBA is also almost equal to the optimal values. What is more, the standard deviation values of IBBA are smaller than those of other algorithms, which means that IBBA can search the optimal value more stably. Thus, it could be concluded that IBBA could provide better capability to solve the zero-one knapsack problem than original BBA and BPSO.
This section briefly presents the reason that the performance of IBBA was better than the other two algorithms:(1)Inspired by several heuristic algorithms, BA combines the merit of previously proposed heuristic algorithms, such as PSO and SA.(2)With the guidance of neighbor bat, the algorithm could produce more diversified solutions.(3)To balance the global search and local search, dynamic inertia weight strategy is employed to control the magnitude of the velocity. Thus, it could avoid premature convergence.(4)The adaptive strategy designed to dynamically control the transformation point of global search and local search promotes the accuracy of results.
In this paper, an improved binary bat algorithm is presented. Simple but effective modifications are employed to enhance the BBA. Then the proposed IBBA is implemented and tested on thirteen well-known benchmark functions and three typical zero-one knapsack problems. The obtained results are compared with those of BBA and BPSO. The experiment results show that the proposed approach could get much better performance than the other two algorithms mentioned above in terms of accuracy of the results.
For further studies, we intend to apply the proposed algorithm to practical applications, such as application switch problem  and task scheduling problem .
Conflicts of Interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
This work was supported by the National High Technology Research and Development Program of China (863 Program) (Grant no. 2015AA015802).
References Z.-L. Gaing, “Discrete particle swarm optimization algorithm for unit commitment,” in Proceedings of the IEEE Power Engineering Society General Meeting, vol. 1, pp. 418–424, Toronto, Canada, 2003. View at Publisher · View at Google ScholarR. Y. Nakamura, L. A. Pereira, K. A. Costa, D. Rodrigues, J. P. Papa, and X. Yang, “Bba: a binary bat algorithm for feature selection,” in Proceedings of the 25th SIBGRAPI Conference on Graphics, Patterns and Images (SIBGRAPI '12), pp. 291–297, Ouro Preto, Brazil, 2012. View at Publisher · View at Google ScholarS. Kashef and H. Nezamabadi-pour, “An advanced ACO algorithm for feature subset selection,” Neurocomputing, vol. 147, no. 1, pp. 271–279, 2015. View at Publisher · View at Google Scholar · View at ScopusH. S. Al-Olimat, M. Alam, R. Green, and J. K. Lee, “Cloudlet scheduling with particle swarm optimization,” in Preceedings of the 5th International Conference on Communication Systems and Network Technologies (CSNT '15), pp. 991–995, April 2015. View at Publisher · View at Google Scholar · View at ScopusL. Xu, K. Wang, Z. Ouyang, and X. Qi, “An improved binary PSO-based task scheduling algorithm in green cloud computing,” in Proceedings of the 9th International Conference on Communications and Networking in China (CHINACOM '14), pp. 126–131, August 2014. View at Publisher · View at Google Scholar · View at ScopusJ. C. Bansal and K. Deep, “A modified binary particle swarm optimization for knapsack problems,” Applied Mathematics and Computation, vol. 218, no. 22, pp. 11042–11061, 2012. View at Publisher · View at Google Scholar · View at MathSciNet · View at ScopusH. Wu, F. Zhang, and R. Zhan, “A binary wolf pack algorithm for solving 0-1 knapsack problem,” Systems Engineering and Electronics, vol. 36, no. 8, pp. 1660–1667, 2014. View at Google ScholarS. Mirjalili, S. M. Mirjalili, and X.-S. Yang, “Binary bat algorithm,” Neural Computing and Applications, vol. 25, no. 3-4, pp. 663–681, 2014. View at Publisher · View at Google Scholar · View at ScopusX.-S. Yang, “A new metaheuristic bat-inspired algorithm,” in Nature Inspired Cooperative Strategies for Optimization (NICSO 2010), J. R. Gonzalez, D. A. Pelta, C. Cruz, G. Terrazas, and N. Krasnogor, Eds., vol. 284 of Studies in Computational Intelligence, pp. 65–74, Springer, Berlin, Germany, 2010. View at Publisher · View at Google ScholarS. P. Brooks and B. J. Morgan, “Optimization using simulated annealing,” The Statistician, vol. 44, no. 2, pp. 241–257, 1995. View at Publisher · View at Google ScholarJ. Kennedy and R. Eberhart, “Particle swarm optimization,” in Proceedings of the IEEE International Conference on Neural Networks (ICNN ’95), vol. 4, pp. 1942–1948, Perth, Western Australia, 1995. View at Publisher · View at Google Scholar · View at ScopusY. Shi and R. Eberhart, “A modified particle swarm optimizer,” in Proceedings of the IEEE International Conference on Evolutionary Computation and IEEE World Congress on Computational Intelligence, pp. 69–73, Anchorage, Ala, USA, 1998. View at Publisher · View at Google ScholarR. C. Eberhart and Y. Shi, “Tracking and optimizing dynamic systems with particle swarms,” in Proceedings of the Congress on Evolutionary Computation, vol. 1, pp. 94–100, Seoul, Republic of Korea, May 2001. View at Publisher · View at Google Scholar · View at ScopusS. U. Khan, S. Yang, L. Wang, and L. Liu, “A modified particle swarm optimization algorithm for global optimizations of inverse problems,” IEEE Transactions on Magnetics, vol. 52, no. 3, 2016. View at Publisher · View at Google Scholar · View at ScopusR. C. Eberhart and Y. Shi, “Comparing inertia weights and constriction factors in particle swarm optimization,” in Proceedings of the 2000 Congress on Evolutionary Computation (CEC' 00), vol. 1, pp. 84–88, La Jolla, Claif, USA, 2000. View at Publisher · View at Google ScholarA. Chatterjee and P. Siarry, “Nonlinear inertia weight variation for dynamic adaptation in particle swarm optimization,” Computers & Operations Research, vol. 33, no. 3, pp. 859–871, 2006. View at Publisher · View at Google ScholarY.-L. Gao, X.-H. An, and J.-M. Liu, “A particle swarm optimization algorithm with logarithm decreasing inertia weight and chaos mutation,” in Proceedings of the 2008 International Conference on Computational Intelligence and Security (CIS '08), pp. 61–65, chn, December 2008. View at Publisher · View at Google Scholar · View at ScopusG. Chen, X. Huang, J. Jia, and Z. Min, “Natural exponential inertia weight strategy in particle swarm optimization,” in Proceedings of the 6th World Congress on Intelligent Control and Automation (WCICA '06), vol. 1, pp. 3672–3675, June 2006. View at Publisher · View at Google Scholar · View at ScopusA. Nickabadi, M. M. Ebadzadeh, and R. Safabakhsh, “A novel particle swarm optimization algorithm with adaptive inertia weight,” Applied Soft Computing Journal, vol. 11, no. 4, pp. 3658–3670, 2011. View at Publisher · View at Google Scholar · View at ScopusZ.-H. Zhan, J. Zhang, Y. Li, and H. S.-H. Chung, “Adaptive particle swarm optimization,” IEEE Transactions on Systems, Man, and Cybernetics Part B: Cybernetics, vol. 39, no. 6, pp. 1362–1381, 2009. View at Publisher · View at Google Scholar · View at ScopusX. Yang, J. Yuan, H. Yuan, and H. Mao, “A modified particle swarm optimizer with dynamic adaptation,” Applied Mathematics & Computation, vol. 189, no. 2, pp. 1205–1213, 2007. View at Publisher · View at Google ScholarS. Yilmaz and E. U. Küçüksille, “A new modification approach on bat algorithm for solving optimization problems,” Applied Soft Computing, vol. 28, pp. 259–275, 2015. View at Publisher · View at Google Scholar · View at ScopusG. Wang, M. Lu, and X. Zhao, “An improved bat algorithm with variable neighborhood search for global optimization,” in Proceedings of the IEEE Congress on Evolutionary Computation (CEC '16), pp. 1773–1778, Vancouver, Canada, 2016. View at Publisher · View at Google ScholarX. Cai, L. Wang, Q. Kang, and Q. Wu, “Bat algorithm with gaussian walk,” International Journal of Bio-Inspired Computation, vol. 6, no. 3, pp. 166–174, 2014. View at Publisher · View at Google Scholar · View at ScopusX. Cai, X. Z. Gao, and Y. Xue, “Improved bat algorithm with optimal forage strategy and random disturbance strategy,” International Journal of Bio-Inspired Computation, vol. 8, no. 4, pp. 205–214, 2016. View at Publisher · View at Google ScholarK. Lei and C. Pu, “Complex optimization problems using highly efficient particle swarm optimizer,” TELKOMNIKA (Telecommunication Computing Electronics and Control), vol. 12, no. 4, pp. 1023–1030, 2014. View at Publisher · View at Google ScholarS. Mirjalili, “Dragonfly algorithm: a new meta-heuristic optimization technique for solving single-objective, discrete, and multi-objective problems,” Neural Computing & Applications, vol. 27, no. 4, pp. 1053–1073, 2016. View at Publisher · View at Google Scholar · View at ScopusJ. Kennedy and R. C. Eberhart, “A discrete binary version of the particle swarm algorithm,” in Proceedings of the IEEE International Conference on Systems, Man, and Cybernetics, Computational Cybernetics and Simulation, vol. 5, pp. 4104–4108, IEEE, Orlando, Fla, USA, October 1997. View at Publisher · View at Google Scholar · View at ScopusF. Wilcoxon, “Individual comparisons by ranking methods,” Biometrics Bulletin, vol. 1, no. 6, pp. 80–83, 1945. View at Google Scholar · View at MathSciNetJ. Derrac, S. García, D. Molina, and F. Herrera, “A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms,” Swarm and Evolutionary Computation, vol. 1, no. 1, pp. 3–18, 2011. View at Publisher · View at Google Scholar · View at ScopusY. Sharafi, M. A. Khanesar, and M. Teshnehlab, “Discrete binary cat swarm optimization algorithm,” in Proceedings of the 3rd IEEE International Conference on Computer, Control and Communication (IC4 '13), pp. 1–6, 2013. View at Publisher · View at Google Scholar · View at ScopusY. Jiang, X. Zeng, and P. Sun, “A rapid application switch technique based on resource cache,” Journal of Network New Media, vol. 2, no. 4, pp. 33–38, 2013. View at Google Scholar