scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Cuckoo search: recent advances and applications

01 Jan 2014-Neural Computing and Applications (Springer London)-Vol. 24, Iss: 1, pp 169-174
TL;DR: Cuckoo search (CS) is a relatively new algorithm, developed by Yang and Deb in 2009, and the same has been found to be efficient in solving global optimization problems.
Abstract: Cuckoo search (CS) is a relatively new algorithm, developed by Yang and Deb in 2009, and the same has been found to be efficient in solving global optimization problems. In this paper, we review the fundamental ideas of cuckoo search and the latest developments as well as its applications. We analyze the algorithm and gain insight into its search mechanisms and find out why it is efficient. We also discuss the essence of algorithms and its link to self-organizing systems, and finally, we propose some important topics for further research.
Citations
More filters
Journal ArticleDOI
TL;DR: A timely review of the bat algorithm and its new variants and a wide range of diverse applications and case studies are reviewed and summarised briefly here.
Abstract: Bat algorithm BA is a bio-inspired algorithm developed by Xin-She Yang in 2010 and BA has been found to be very efficient. As a result, the literature has expanded significantly in the last three years. This paper provides a timely review of the bat algorithm and its new variants. A wide range of diverse applications and case studies are also reviewed and summarised briefly here. In addition, we also discuss the essence of an algorithm and the links between algorithms and self-organisation. Further research topics are also discussed.

791 citations

Journal ArticleDOI
TL;DR: A comparative study with five other metaheuristic algorithms through thirty-eight benchmark problems is carried out, and the results clearly exhibit the capability of the MBO method toward finding the enhanced function values on most of the benchmark problems with respect to the other five algorithms.
Abstract: In nature, the eastern North American monarch population is known for its southward migration during the late summer/autumn from the northern USA and southern Canada to Mexico, covering thousands of miles. By simplifying and idealizing the migration of monarch butterflies, a new kind of nature-inspired metaheuristic algorithm, called monarch butterfly optimization (MBO), a first of its kind, is proposed in this paper. In MBO, all the monarch butterfly individuals are located in two distinct lands, viz. southern Canada and the northern USA (Land 1) and Mexico (Land 2). Accordingly, the positions of the monarch butterflies are updated in two ways. Firstly, the offsprings are generated (position updating) by migration operator, which can be adjusted by the migration ratio. It is followed by tuning the positions for other butterflies by means of butterfly adjusting operator. In order to keep the population unchanged and minimize fitness evaluations, the sum of the newly generated butterflies in these two ways remains equal to the original population. In order to demonstrate the superior performance of the MBO algorithm, a comparative study with five other metaheuristic algorithms through thirty-eight benchmark problems is carried out. The results clearly exhibit the capability of the MBO method toward finding the enhanced function values on most of the benchmark problems with respect to the other five algorithms. Note that the source codes of the proposed MBO algorithm are publicly available at GitHub ( https://github.com/ggw0122/Monarch-Butterfly-Optimization , C++/MATLAB) and MATLAB Central ( http://www.mathworks.com/matlabcentral/fileexchange/50828-monarch-butterfly-optimization , MATLAB).

778 citations

Proceedings ArticleDOI
07 Dec 2015
TL;DR: A new kind of swarm-based metaheuristic search method, called Elephant Herding Optimization (EHO), is proposed for solving optimization tasks, inspired by the herding behavior of elephant group.
Abstract: In this paper, a new kind of swarm-based metaheuristic search method, called Elephant Herding Optimization (EHO), is proposed for solving optimization tasks The EHO method is inspired by the herding behavior of elephant group In nature, the elephants belonging to different clans live together under the leadership of a matriarch, and the male elephants will leave their family group when they grow up These two behaviors can be modelled into two following operators: clan updating operator and separating operator In EHO, the elephants in each clan are updated by its current position and matriarch through clan updating operator It is followed by the implementation of the separating operator which can enhance the population diversity at the later search phase To demonstrate its effectiveness, EHO is benchmarked by fifteen test cases comparing with BBO, DE and GA The results show that EHO can find the better values on most benchmark problems than those three metaheuristic algorithms

548 citations

Journal ArticleDOI
TL;DR: A survey of metaheuristic research in literature consisting of 1222 publications from year 1983 to 2016 is performed to highlight potential open questions and critical issues raised in literature and provides guidance for future research to be conducted more meaningfully.
Abstract: Because of successful implementations and high intensity, metaheuristic research has been extensively reported in literature, which covers algorithms, applications, comparisons, and analysis. Though, little has been evidenced on insightful analysis of metaheuristic performance issues, and it is still a “black box” that why certain metaheuristics perform better on specific optimization problems and not as good on others. The performance related analyses performed on algorithms are mostly quantitative via performance validation metrics like mean error, standard deviation, and co-relations have been used. Moreover, the performance tests are often performed on specific benchmark functions—few studies are those which involve real data from scientific or engineering optimization problems. In order to draw a comprehensive picture of metaheuristic research, this paper performs a survey of metaheuristic research in literature which consists of 1222 publications from year 1983 to 2016 (33 years). Based on the collected evidence, this paper addresses four dimensions of metaheuristic research: introduction of new algorithms, modifications and hybrids, comparisons and analysis, and research gaps and future directions. The objective is to highlight potential open questions and critical issues raised in literature. The work provides guidance for future research to be conducted more meaningfully that can serve for the good of this area of research.

467 citations

Journal ArticleDOI
TL;DR: In this survey, fourteen new and outstanding metaheuristics that have been introduced for the last twenty years other than the classical ones such as genetic, particle swarm, and tabu search are distinguished.

450 citations

References
More filters
Proceedings ArticleDOI
06 Aug 2002
TL;DR: A concept for the optimization of nonlinear functions using particle swarm methodology is introduced, and the evolution of several paradigms is outlined, and an implementation of one of the paradigm is discussed.
Abstract: A concept for the optimization of nonlinear functions using particle swarm methodology is introduced. The evolution of several paradigms is outlined, and an implementation of one of the paradigms is discussed. Benchmark testing of the paradigm is described, and applications, including nonlinear function optimization and neural network training, are proposed. The relationships between particle swarm optimization and both artificial life and genetic algorithms are described.

35,104 citations

Journal ArticleDOI
TL;DR: A framework is developed to explore the connection between effective optimization algorithms and the problems they are solving and a number of "no free lunch" (NFL) theorems are presented which establish that for any algorithm, any elevated performance over one class of problems is offset by performance over another class.
Abstract: A framework is developed to explore the connection between effective optimization algorithms and the problems they are solving. A number of "no free lunch" (NFL) theorems are presented which establish that for any algorithm, any elevated performance over one class of problems is offset by performance over another class. These theorems result in a geometric interpretation of what it means for an algorithm to be well suited to an optimization problem. Applications of the NFL theorems to information-theoretic aspects of optimization and benchmark measures of performance are also presented. Other issues addressed include time-varying optimization problems and a priori "head-to-head" minimax distinctions between optimization algorithms, distinctions that result despite the NFL theorems' enforcing of a type of uniformity over all algorithms.

10,771 citations

Journal ArticleDOI
TL;DR: This paper analyzes a particle's trajectory as it moves in discrete time, then progresses to the view of it in continuous time, leading to a generalized model of the algorithm, containing a set of coefficients to control the system's convergence tendencies.
Abstract: The particle swarm is an algorithm for finding optimal regions of complex search spaces through the interaction of individuals in a population of particles. This paper analyzes a particle's trajectory as it moves in discrete time (the algebraic view), then progresses to the view of it in continuous time (the analytical view). A five-dimensional depiction is developed, which describes the system completely. These analyses lead to a generalized model of the algorithm, containing a set of coefficients to control the system's convergence tendencies. Some results of the particle swarm optimizer, implementing modifications derived from the analysis, suggest methods for altering the original algorithm in ways that eliminate problems and increase the ability of the particle swarm to find optima of some well-studied test functions.

8,287 citations

Proceedings ArticleDOI
01 Dec 2009
TL;DR: A new meta-heuristic algorithm, called Cuckoo Search (CS), is formulated, based on the obligate brood parasitic behaviour of some cuckoo species in combination with the Lévy flight behaviour ofSome birds and fruit flies, for solving optimization problems.
Abstract: In this paper, we intend to formulate a new meta-heuristic algorithm, called Cuckoo Search (CS), for solving optimization problems. This algorithm is based on the obligate brood parasitic behaviour of some cuckoo species in combination with the Levy flight behaviour of some birds and fruit flies. We validate the proposed algorithm against test functions and then compare its performance with those of genetic algorithms and particle swarm optimization. Finally, we discuss the implication of the results and suggestion for further research.

5,521 citations

Book ChapterDOI
26 Oct 2009
TL;DR: In this article, a new Firefly Algorithm (FA) was proposed for multimodal optimization applications. And the proposed FA was compared with other metaheuristic algorithms such as particle swarm optimization (PSO).
Abstract: Nature-inspired algorithms are among the most powerful algorithms for optimization. This paper intends to provide a detailed description of a new Firefly Algorithm (FA) for multimodal optimization applications. We will compare the proposed firefly algorithm with other metaheuristic algorithms such as particle swarm optimization (PSO). Simulations and results indicate that the proposed firefly algorithm is superior to existing metaheuristic algorithms. Finally we will discuss its applications and implications for further research.

3,436 citations