scispace - formally typeset
Search or ask a question

Showing papers on "Metaheuristic published in 2009"


Proceedings ArticleDOI
01 Dec 2009
TL;DR: A new meta-heuristic algorithm, called Cuckoo Search (CS), is formulated, based on the obligate brood parasitic behaviour of some cuckoo species in combination with the Lévy flight behaviour ofSome birds and fruit flies, for solving optimization problems.
Abstract: In this paper, we intend to formulate a new meta-heuristic algorithm, called Cuckoo Search (CS), for solving optimization problems. This algorithm is based on the obligate brood parasitic behaviour of some cuckoo species in combination with the Levy flight behaviour of some birds and fruit flies. We validate the proposed algorithm against test functions and then compare its performance with those of genetic algorithms and particle swarm optimization. Finally, we discuss the implication of the results and suggestion for further research.

5,521 citations


Journal ArticleDOI
TL;DR: A new optimization algorithm based on the law of gravity and mass interactions is introduced and the obtained results confirm the high performance of the proposed method in solving various nonlinear functions.

5,501 citations


Book ChapterDOI
26 Oct 2009
TL;DR: In this article, a new Firefly Algorithm (FA) was proposed for multimodal optimization applications. And the proposed FA was compared with other metaheuristic algorithms such as particle swarm optimization (PSO).
Abstract: Nature-inspired algorithms are among the most powerful algorithms for optimization. This paper intends to provide a detailed description of a new Firefly Algorithm (FA) for multimodal optimization applications. We will compare the proposed firefly algorithm with other metaheuristic algorithms such as particle swarm optimization (PSO). Simulations and results indicate that the proposed firefly algorithm is superior to existing metaheuristic algorithms. Finally we will discuss its applications and implications for further research.

3,436 citations


Book
22 Jun 2009
TL;DR: This book provides a complete background on metaheuristics and shows readers how to design and implement efficient algorithms to solve complex optimization problems across a diverse range of applications, from networking and bioinformatics to engineering design, routing, and scheduling.
Abstract: A unified view of metaheuristics This book provides a complete background on metaheuristics and shows readers how to design and implement efficient algorithms to solve complex optimization problems across a diverse range of applications, from networking and bioinformatics to engineering design, routing, and scheduling. It presents the main design questions for all families of metaheuristics and clearly illustrates how to implement the algorithms under a software framework to reuse both the design and code. Throughout the book, the key search components of metaheuristics are considered as a toolbox for: Designing efficient metaheuristics (e.g. local search, tabu search, simulated annealing, evolutionary algorithms, particle swarm optimization, scatter search, ant colonies, bee colonies, artificial immune systems) for optimization problems Designing efficient metaheuristics for multi-objective optimization problems Designing hybrid, parallel, and distributed metaheuristics Implementing metaheuristics on sequential and parallel machines Using many case studies and treating design and implementation independently, this book gives readers the skills necessary to solve large-scale optimization problems quickly and efficiently. It is a valuable reference for practicing engineers and researchers from diverse areas dealing with optimization or machine learning; and graduate students in computer science, operations research, control, engineering, business and management, and applied mathematics.

2,735 citations


Book ChapterDOI
21 Apr 2009
TL;DR: Ant Colony Optimization (ACO) is a stochastic local search method that has been inspired by the pheromone trail laying and following behavior of some ant species as discussed by the authors.
Abstract: Ant Colony Optimization (ACO) is a stochastic local search method that has been inspired by the pheromone trail laying and following behavior of some ant species [1]. Artificial ants in ACO essentially are randomized construction procedures that generate solutions based on (artificial) pheromone trails and heuristic information that are associated to solution components. Since the first ACO algorithm has been proposed in 1991, this algorithmic method has attracted a large number of researchers and in the meantime it has reached a significant level of maturity. In fact, ACO is now a well-established search technique for tackling a wide variety of computationally hard problems.

2,424 citations


Journal ArticleDOI
TL;DR: A family of improved variants of the DE/target-to-best/1/bin scheme, which utilizes the concept of the neighborhood of each population member, and is shown to be statistically significantly better than or at least comparable to several existing DE variants as well as a few other significant evolutionary computing techniques over a test suite of 24 benchmark functions.
Abstract: Differential evolution (DE) is well known as a simple and efficient scheme for global optimization over continuous spaces. It has reportedly outperformed a few evolutionary algorithms (EAs) and other search heuristics like the particle swarm optimization (PSO) when tested over both benchmark and real-world problems. DE, however, is not completely free from the problems of slow and/or premature convergence. This paper describes a family of improved variants of the DE/target-to-best/1/bin scheme, which utilizes the concept of the neighborhood of each population member. The idea of small neighborhoods, defined over the index-graph of parameter vectors, draws inspiration from the community of the PSO algorithms. The proposed schemes balance the exploration and exploitation abilities of DE without imposing serious additional burdens in terms of function evaluations. They are shown to be statistically significantly better than or at least comparable to several existing DE variants as well as a few other significant evolutionary computing techniques over a test suite of 24 benchmark functions. The paper also investigates the applications of the new DE variants to two real-life problems concerning parameter estimation for frequency modulated sound waves and spread spectrum radar poly-phase code design.

1,086 citations


Gilbert Laporte1
01 Aug 2009
TL;DR: The Vehicle Routing Problem was introduced 50 years ago by Dantzig and Ramser under the title “The Truck Dispatching Problem” and has given rise to major developments in the fields of exact algorithms and heuristics.
Abstract: The Vehicle Routing Problem (VRP) was introduced 50 years ago by Dantzig and Ramser under the title “The Truck Dispatching Problem.” The study of the VRP has given rise to major developments in the fields of exact algorithms and heuristics. In particular, highly sophisticated exact mathematical programming decomposition algorithms and powerful metaheuristics for the VRP have been put forward in recent years. The purpose of this article is to provide a brief account of this development.

762 citations


Journal ArticleDOI
TL;DR: In this paper metaheuristics such as Ant Colony Optimization, Evolutionary Computation, Simulated Annealing, Tabu Search and others are introduced, and their applications to the class of Stochastic Combinatorial Optimization Problems (SCOPs) is thoroughly reviewed.
Abstract: Metaheuristics are general algorithmic frameworks, often nature-inspired, designed to solve complex optimization problems, and they are a growing research area since a few decades. In recent years, metaheuristics are emerging as successful alternatives to more classical approaches also for solving optimization problems that include in their mathematical formulation uncertain, stochastic, and dynamic information. In this paper metaheuristics such as Ant Colony Optimization, Evolutionary Computation, Simulated Annealing, Tabu Search and others are introduced, and their applications to the class of Stochastic Combinatorial Optimization Problems (SCOPs) is thoroughly reviewed. Issues common to all metaheuristics, open problems, and possible directions of research are proposed and discussed. In this survey, the reader familiar to metaheuristics finds also pointers to classical algorithmic approaches to optimization under uncertainty, and useful informations to start working on this problem domain, while the reader new to metaheuristics should find a good tutorial in those metaheuristics that are currently being applied to optimization under uncertainty, and motivations for interest in this field.

638 citations


Journal ArticleDOI
TL;DR: This work presents a survey of the algorithms described based on the intelligence in bee swarms and their applications, and presents a list of winners and losers.
Abstract: Swarm intelligence is an emerging area in the field of optimization and researchers have developed various algorithms by modeling the behaviors of different swarm of animals and insects such as ants, termites, bees, birds, fishes. In 1990s, Ant Colony Optimization based on ant swarm and Particle Swarm Optimization based on bird flocks and fish schools have been introduced and they have been applied to solve optimization problems in various areas within a time of two decade. However, the intelligent behaviors of bee swarm have inspired the researchers especially during the last decade to develop new algorithms. This work presents a survey of the algorithms described based on the intelligence in bee swarms and their applications.

624 citations


Proceedings ArticleDOI
15 May 2009
TL;DR: A new multi-objective particle swarm optimization algorithm characterized by the use of a strategy to limit the velocity of the particles, called Speed-constrained Multi-Objective PSO (SMPSO), which allows to produce new effective particle positions in those cases in which the velocity becomes too high.
Abstract: In this work, we present a new multi-objective particle swarm optimization algorithm (PSO) characterized by the use of a strategy to limit the velocity of the particles. The proposed approach, called Speed-constrained Multi-objective PSO (SMPSO) allows to produce new effective particle positions in those cases in which the velocity becomes too high. Other features of SMPSO include the use of polynomial mutation as a turbulence factor and an external archive to store the non-dominated solutions found during the search. Our proposed approach is compared with respect to five multi-objective metaheuristics representative of the state-of-the-art in the area. For the comparison, two different criteria are adopted: the quality of the resulting approximation sets and the convergence speed to the Pareto front. The experiments carried out indicate that SMPSO obtains remarkable results in terms of both, accuracy and speed.

563 citations


Book
14 Aug 2009
TL;DR: In this article, a cost minimization model for the design of water distribution networks is presented, which uses a recently developed harmony search optimization algorithm while satisfying all the design constraints, including pipe diameters in a water distribution network.
Abstract: This study presents a cost minimization model for the design of water distribution networks. The model uses a recently developed harmony search optimization algorithm while satisfying all the design constraints. The harmony search algorithm mimics a jazz improvisation process in order to find better design solutions, in this case pipe diameters in a water distribution network. The model also interfaces with a popular hydraulic simulator, EPANET, to check the hydraulic constraints. If the design solution vector violates the hydraulic constraints, the amount of violation is considered in the cost function as a penalty. The model was applied to five water distribution networks, and obtained designs that were either the same or cost 0.28–10.26% less than those of competitive meta-heuristic algorithms, such as the genetic algorithm, simulated annealing and tabu search under similar or less favorable conditions. The results show that the harmony search-based model is suitable for water network design.

Book ChapterDOI
TL;DR: This first chapter intends to review and analyze the powerful new Harmony Search algorithm in the context of metaheuristic algorithms, and tries to identify the characteristics of meta heuristics and analyze why HS is a good meta heuristic algorithm.
Abstract: This first chapter intends to review and analyze the powerful new Harmony Search (HS) algorithm in the context of metaheuristic algorithms. We will first outline the fundamental steps of HS, and show how it works. We then try to identify the characteristics of metaheuristics and analyze why HS is a good metaheuristic algorithm. We then review briefly other popular metaheuristics such as particle swarm optimization so as to find their similarities and differences with HS. Finally, we will discuss the ways to improve and develop new variants of HS, and make suggestions for further research including open questions.

Journal ArticleDOI
TL;DR: This paper presents an overview of the research progress in deterministic global optimization during the last decade (1998-2008).
Abstract: This paper presents an overview of the research progress in deterministic global optimization during the last decade (1998---2008). It covers the areas of twice continuously differentiable nonlinear optimization, mixed-integer nonlinear optimization, optimization with differential-algebraic models, semi-infinite programming, optimization with grey box/nonfactorable models, and bilevel nonlinear optimization.

Journal ArticleDOI
TL;DR: In this paper, a heuristic particle swarm ant colony optimization (HPSACO) is presented for optimum design of trusses, which is based on the particle swarm optimizer with passive congregation (PSOPC), ant colony optimizer and harmony search scheme.

Journal ArticleDOI
TL;DR: It has been detected that coupling emergent results in different areas, like those of PSO and complex dynamics, can improve the quality of results in some optimization problems.
Abstract: This paper proposes new particle swarm optimization (PSO) methods that use chaotic maps for parameter adaptation. This has been done by using of chaotic number generators each time a random number is needed by the classical PSO algorithm. Twelve chaos-embedded PSO methods have been proposed and eight chaotic maps have been analyzed in the benchmark functions. It has been detected that coupling emergent results in different areas, like those of PSO and complex dynamics, can improve the quality of results in some optimization problems. It has been also shown that, some of the proposed methods have somewhat increased the solution quality, that is in some cases they improved the global searching capability by escaping the local solutions.

Journal ArticleDOI
TL;DR: The computational results have proved that the proposed hybrid algorithm is an efficient and effective approach to solve the multi-objective FJSP, especially for the problems on a large scale.

Journal ArticleDOI
Alok Singh1
01 Mar 2009
TL;DR: The proposed ABC algorithm is a new metaheuristic approach inspired by intelligent foraging behavior of honeybee swarm that obtained better quality solutions in shorter time and was compared against the best approaches reported in the literature.
Abstract: Given an undirected, connected, weighted graph, the leaf-constrained minimum spanning tree (LCMST) problem seeks on this graph a spanning tree of minimum weight among all the spanning trees of the graph that have at least @? leaves. In this paper, we have proposed an artificial bee colony (ABC) algorithm for the LCMST problem. The ABC algorithm is a new metaheuristic approach inspired by intelligent foraging behavior of honeybee swarm. We have compared the performance of our ABC approach against the best approaches reported in the literature. Computational results demonstrate the superiority of the new ABC approach over all the other approaches. The new approach obtained better quality solutions in shorter time.

Journal ArticleDOI
TL;DR: A variety of metaheuristic search techniques are found to be applicable for non-functional testing including simulated annealing, tabu search, genetic algorithms, ant colony methods, grammatical evolution, genetic programming and swarm intelligence methods.
Abstract: Search-based software testing is the application of metaheuristic search techniques to generate software tests. The test adequacy criterion is transformed into a fitness function and a set of solutions in the search space are evaluated with respect to the fitness function using a metaheuristic search technique. The application of metaheuristic search techniques for testing is promising due to the fact that exhaustive testing is infeasible considering the size and complexity of software under test. Search-based software testing has been applied across the spectrum of test case design methods; this includes white-box (structural), black-box (functional) and grey-box (combination of structural and functional) testing. In addition, metaheuristic search techniques have also been applied to test non-functional properties. The overall objective of undertaking this systematic review is to examine existing work into non-functional search-based software testing (NFSBST). We are interested in types of non-functional testing targeted using metaheuristic search techniques, different fitness functions used in different types of search-based non-functional testing and challenges in the application of these techniques. The systematic review is based on a comprehensive set of 35 articles obtained after a multi-stage selection process and have been published in the time span 1996-2007. The results of the review show that metaheuristic search techniques have been applied for non-functional testing of execution time, quality of service, security, usability and safety. A variety of metaheuristic search techniques are found to be applicable for non-functional testing including simulated annealing, tabu search, genetic algorithms, ant colony methods, grammatical evolution, genetic programming (and its variants including linear genetic programming) and swarm intelligence methods. The review reports on different fitness functions used to guide the search for each of the categories of execution time, safety, usability, quality of service and security; along with a discussion of possible challenges in the application of metaheuristic search techniques.

Journal ArticleDOI
TL;DR: Experimental results demonstrate the efficacy of the proposed glowworm based algorithm in capturing multiple optima of a series of standard multimodal test functions and more complex ones, such as stair-case and multiple-plateau functions.
Abstract: This paper presents glowworm swarm optimization (GSO), a novel algorithm for the simultaneous computation of multiple optima of multimodal functions. The algorithm shares a few features with some better known swarm intelligence based optimization algorithms, such as ant colony optimization and particle swarm optimization, but with several significant differences. The agents in GSO are thought of as glowworms that carry a luminescence quantity called luciferin along with them. The glowworms encode the fitness of their current locations, evaluated using the objective function, into a luciferin value that they broadcast to their neighbors. The glowworm identifies its neighbors and computes its movements by exploiting an adaptive neighborhood, which is bounded above by its sensor range. Each glowworm selects, using a probabilistic mechanism, a neighbor that has a luciferin value higher than its own and moves toward it. These movements—based only on local information and selective neighbor interactions—enable the swarm of glowworms to partition into disjoint subgroups that converge on multiple optima of a given multimodal function. We provide some theoretical results related to the luciferin update mechanism in order to prove the bounded nature and convergence of luciferin levels of the glowworms. Experimental results demonstrate the efficacy of the proposed glowworm based algorithm in capturing multiple optima of a series of standard multimodal test functions and more complex ones, such as stair-case and multiple-plateau functions. We also report the results of tests in higher dimensional spaces with a large number of peaks. We address the parameter selection problem by conducting experiments to show that only two parameters need to be selected by the user. Finally, we provide some comparisons of GSO with PSO and an experimental comparison with Niche-PSO, a PSO variant that is designed for the simultaneous computation of multiple optima.

Journal ArticleDOI
TL;DR: An improved ant colony optimization (IACO) is proposed, which possesses a new strategy to update the increased pheromone, called ant-weight strategy, and a mutation operation, to solve VRP.

Book ChapterDOI
04 Oct 2009
TL;DR: The paper provides an insight into the improved novel metaheuristics of the Firefly Algorithm for constrained continuous optimization tasks and some suggestions for extending the simple scheme of the technique under consideration are presented.
Abstract: The paper provides an insight into the improved novel metaheuristics of the Firefly Algorithm for constrained continuous optimization tasks. The presented technique is inspired by social behavior of fireflies and the phenomenon of bioluminescent communication. The first part of the paper is devoted to the detailed description of the existing algorithm. Then some suggestions for extending the simple scheme of the technique under consideration are presented. Subsequent sections concentrate on the performed experimental parameter studies and a comparison with existing Particle Swarm Optimization strategy based on existing benchmark instances. Finally some concluding remarks on possible algorithm extensions are given, as well as some properties of the presented approach and comments on its performance in the constrained continuous optimization tasks.

Journal ArticleDOI
TL;DR: A Hybrid Big Bang-Big Crunch (HBB-BC) optimization algorithm is employed for optimal design of truss structures and numerical results demonstrate the efficiency and robustness of the H BB-BC method compared to other heuristic algorithms.

Journal ArticleDOI
TL;DR: A real-value version of particle swarm optimization (PSO) algorithm for solving the vehicle routing problem with simultaneous pickup and delivery (VRPSPD) and the mechanism of the PSO for solving VRPSPD is explained and demonstrated.

Journal ArticleDOI
TL;DR: This paper presents an evolutionary algorithm, entitled A Multialgorithm Genetically Adaptive Method for Single Objective Optimization (AMALGAM-SO), that implements this concept of self adaptive multimethod search and implements a self-adaptive learning strategy to automatically tune the number of offspring these three individual algorithms are allowed to contribute during each generation.
Abstract: Many different algorithms have been developed in the last few decades for solving complex real-world search and optimization problems. The main focus in this research has been on the development of a single universal genetic operator for population evolution that is always efficient for a diverse set of optimization problems. In this paper, we argue that significant advances to the field of evolutionary computation can be made if we embrace a concept of self-adaptive multimethod optimization in which multiple different search algorithms are run concurrently, and learn from each other through information exchange using a common population of points. We present an evolutionary algorithm, entitled A Multialgorithm Genetically Adaptive Method for Single Objective Optimization (AMALGAM-SO), that implements this concept of self adaptive multimethod search. This method simultaneously merges the strengths of the covariance matrix adaptation (CMA) evolution strategy, genetic algorithm (GA), and particle swarm optimizer (PSO) for population evolution and implements a self-adaptive learning strategy to automatically tune the number of offspring these three individual algorithms are allowed to contribute during each generation. Benchmark results in 10, 30, and 50 dimensions using synthetic functions from the special session on real-parameter optimization of CEC 2005 show that AMALGAM-SO obtains similar efficiencies as existing algorithms on relatively simple unimodal problems, but is superior for more complex higher dimensional multimodal optimization problems. The new search method scales well with increasing number of dimensions, converges in the close proximity of the global minimum for functions with noise induced multimodality, and is designed to take full advantage of the power of distributed computer networks.

Book
12 May 2009
TL;DR: This book especially focuses on a music-inspired metaheuristic algorithm, harmony search, where variables in the harmony search algorithm have random values or previously-memorized good values in order to find optimal solution.
Abstract: Calculus has been used in solving many scientific and engineering problems. For optimization problems, however, the differential calculus technique sometimes has a drawback when the objective function is step-wise, discontinuous, or multi-modal, or when decision variables are discrete rather than continuous. Thus, researchers have recently turned their interests into metaheuristic algorithms that have been inspired by natural phenomena such as evolution, animal behavior, or metallic annealing. This book especially focuses on a music-inspired metaheuristic algorithm, harmony search. Interestingly, there exists an analogy between music and optimization: each musical instrument corresponds to each decision variable; musical note corresponds to variable value; and, harmony corresponds to solution vector. Just like musicians in Jazz improvisation play notes randomly or based on experiences in order to find fantastic harmony, variables in the harmony search algorithm have random values or previously-memorized good values in order to find optimal solution.

Book
08 Apr 2009
TL;DR: By adopting a machine learning perspective, the author gives a formal definition of the tuning problem, develops a generic algorithm for tuning meta heuristics, and defines an appropriate experimental methodology for assessing the performance of metaheuristics.
Abstract: The importance of tuning metaheuristics is widely acknowledged in scientific literature. However, there is very little dedicated research on the subject. Typically, scientists and practitioners tune metaheuristics by hand, guided only by their experience and by some rules of thumb. Tuning metaheuristics is often considered to be more of an art than a science. This book lays the foundations for a scientific approach to tuning metaheuristics. The fundamental intuition that underlies Birattari's approach is that the tuning problem has much in common with the problems that are typically faced in machine learning. By adopting a machine learning perspective, the author gives a formal definition of the tuning problem, develops a generic algorithm for tuning metaheuristics, and defines an appropriate experimental methodology for assessing the performance of metaheuristics.

Journal ArticleDOI
TL;DR: This paper proposes embedding constraint handling methods, which include the gradient repair method and constraint fitness priority-based ranking method, in NM-PSO as a special operator to deal with satisfying constraints.
Abstract: Constrained optimization problems are very important in that they frequently appear in the real world. A constrained optimization problem consists of the optimization of a function subject to constraints, in which both the function and constraints may be nonlinear. Constraint handling is one of the major concerns when solving constrained optimization problems by hybrid Nelder-Mead simplex search method and particle swarm optimization, denoted as NM-PSO. This paper proposes embedding constraint handling methods, which include the gradient repair method and constraint fitness priority-based ranking method, in NM-PSO as a special operator to deal with satisfying constraints. Experiments using three benchmark function and three engineering design problems are presented and compared with the best known solutions reported in the literature. The comparison results with other evolutionary optimization methods demonstrate that NM-PSO with the embedded constraint operator proves to be extremely effective and efficient at locating optimal solutions.

Journal ArticleDOI
TL;DR: A simple, fast and effective iterated local search meta-heuristic to solve the TOPTW and an insert step is combined with a shake step to escape from local optima, produces a heuristic that performs very well on a large and diverse set of instances.

Journal ArticleDOI
TL;DR: It is shown that slight changes of the proposed VNS procedure is also competitive for the Periodic Traveling Salesman Problem (PTSP), and even outperforms existing solution procedures proposed in the literature.

Proceedings ArticleDOI
08 Jul 2009
TL;DR: Compared with other PSO algorithms, the comparisons show that OLPSO significantly improves the performance of PSO, offering faster global convergence, higher solution quality, and stronger robustness.
Abstract: This paper proposes an orthogonal learning particle swarm optimization (OLPSO) by designing an orthogonal learning (OL) strategy through the orthogonal experimental design (OED) method. The OL strategy takes the dimensions of the problem as the orthogonal experimental factors. The levels of each dimension (factor) are the two choices of the personal best position and the neighborhood's best position. By orthogonally combining the two learning exemplars, the useful information can be discovered, preserved and utilized to construct an efficient exemplar to guide the particle to fly in a more promising direction towards the global optimum. The effectiveness and efficiency of the OL strategy is demonstrated on a set of benchmark functions by comparing the PSOs with and without OL strategy. The OL strategy improves the PSO algorithm in terms of higher quality solution and faster convergence speed.