scispace - formally typeset
Search or ask a question

Showing papers on "Heuristic (computer science) published in 2007"


Journal ArticleDOI
TL;DR: The impacts of constant parameters on harmony search algorithm are discussed and a strategy for tuning these parameters is presented and the proposed algorithm can find better solutions when compared to HS and other heuristic or deterministic methods.

1,782 citations


Journal ArticleDOI
TL;DR: A unified heuristic which is able to solve five different variants of the vehicle routing problem and shown promising results for a large class of vehicle routing problems with backhauls as demonstrated in Ropke and Pisinger.

1,282 citations


Proceedings ArticleDOI
01 Apr 2007
TL;DR: A standard algorithm is defined here which is designed to be a straightforward extension of the original algorithm while taking into account more recent developments that can be expected to improve performance on standard measures.
Abstract: Particle swarm optimization has become a common heuristic technique in the optimization community, with many researchers exploring the concepts, issues, and applications of the algorithm. In spite of this attention, there has as yet been no standard definition representing exactly what is involved in modern implementations of the technique. A standard is defined here which is designed to be a straightforward extension of the original algorithm while taking into account more recent developments that can be expected to improve performance on standard measures. This standard algorithm is intended for use both as a baseline for performance testing of improvements to the technique, as well as to represent PSO to the wider optimization community

1,269 citations


Journal ArticleDOI
TL;DR: A broker-based architecture is designed to facilitate the selection of QoS-based services and efficient heuristic algorithms for service processes of different composition structures are presented.
Abstract: Service-Oriented Architecture (SOA) provides a flexible framework for service composition Using standard-based protocols (such as SOAP and WSDL), composite services can be constructed by integrating atomic services developed independently Algorithms are needed to select service components with various QoS levels according to some application-dependent performance requirements We design a broker-based architecture to facilitate the selection of QoS-based services The objective of service selection is to maximize an application-specific utility function under the end-to-end QoS constraints The problem is modeled in two ways: the combinatorial model and the graph model The combinatorial model defines the problem as a multidimension multichoice 0-1 knapsack problem (MMKP) The graph model defines the problem as a multiconstraint optimal path (MCOP) problem Efficient heuristic algorithms for service processes of different composition structures are presented in this article and their performances are studied by simulations We also compare the pros and cons between the two models

1,225 citations


Journal ArticleDOI
TL;DR: This work presents a new iterated greedy algorithm that applies two phases iteratively, named destruction, were some jobs are eliminated from the incumbent solution, and construction, where the eliminated jobs are reinserted into the sequence using the well known NEH construction heuristic.

923 citations


Journal ArticleDOI
TL;DR: This paper proposes a classification scheme and looks at a number of problem variants in location-routing: a relatively new branch of locational analysis that takes into account vehicle routing aspects.

907 citations


Journal ArticleDOI
TL;DR: This paper presents some of the most important QAP formulations and classify them according to their mathematical sources and gives a detailed discussion of the progress made in both exact and heuristic solution methods, including those formulated according to metaheuristic strategies.

648 citations


Journal ArticleDOI
TL;DR: The algorithm described here, called OptQuest/NLP or OQNLP, is a heuristic designed to find global optima for pure and mixed integer nonlinear problems with many constraints and variables, where all problem functions are differentiable with respect to the continuous variables.
Abstract: The algorithm described here, called OptQuest/NLP or OQNLP, is a heuristic designed to find global optima for pure and mixed integer nonlinear problems with many constraints and variables, where all problem functions are differentiable with respect to the continuous variables. It uses OptQuest, a commercial implementation of scatter search developed by OptTek Systems, Inc., to provide starting points for any gradient-based local solver for nonlinear programming (NLP) problems. This solver seeks a local solution from a subset of these points, holding discrete variables fixed. The procedure is motivated by our desire to combine the superior accuracy and feasibility-seeking behavior of gradient-based local NLP solvers with the global optimization abilities of OptQuest. Computational results include 155 smooth NLP and mixed integer nonlinear program (MINLP) problems due to Floudas et al. (1999), most with both linear and nonlinear constraints, coded in the GAMS modeling language. Some are quite large for global optimization, with over 100 variables and 100 constraints. Global solutions to almost all problems are found in a small number of local solver calls, often one or two.

631 citations


Journal ArticleDOI
01 Feb 2007
TL;DR: This paper proposes an effective particle swarm optimization (PSO)-based memetic algorithm (MA) for the permutation flow shop scheduling problem (PFSSP) with the objective to minimize the maximum completion time, which is a typical non-deterministic polynomial-time (NP) hard combinatorial optimization problem.
Abstract: This paper proposes an effective particle swarm optimization (PSO)-based memetic algorithm (MA) for the permutation flow shop scheduling problem (PFSSP) with the objective to minimize the maximum completion time, which is a typical non-deterministic polynomial-time (NP) hard combinatorial optimization problem. In the proposed PSO-based MA (PSOMA), both PSO-based searching operators and some special local searching operators are designed to balance the exploration and exploitation abilities. In particular, the PSOMA applies the evolutionary searching mechanism of PSO, which is characterized by individual improvement, population cooperation, and competition to effectively perform exploration. On the other hand, the PSOMA utilizes several adaptive local searches to perform exploitation. First, to make PSO suitable for solving PFSSP, a ranked-order value rule based on random key representation is presented to convert the continuous position values of particles to job permutations. Second, to generate an initial swarm with certain quality and diversity, the famous Nawaz-Enscore-Ham (NEH) heuristic is incorporated into the initialization of population. Third, to balance the exploration and exploitation abilities, after the standard PSO-based searching operation, a new local search technique named NEH_1 insertion is probabilistically applied to some good particles selected by using a roulette wheel mechanism with a specified probability. Fourth, to enrich the searching behaviors and to avoid premature convergence, a simulated annealing (SA)-based local search with multiple different neighborhoods is designed and incorporated into the PSOMA. Meanwhile, an effective adaptive meta-Lamarckian learning strategy is employed to decide which neighborhood to be used in SA-based local search. Finally, to further enhance the exploitation ability, a pairwise-based local search is applied after the SA-based search. Simulation results based on benchmarks demonstrate the effectiveness of the PSOMA. Additionally, the effects of some parameters on optimization performances are also discussed

451 citations


Journal ArticleDOI
TL;DR: This paper presents a mixed integer nonlinear programming model for the design of a dynamic integrated distribution network to account for the integrated aspect of optimizing the forward and return network simultaneously.

420 citations


Proceedings Article
23 Sep 2007
TL;DR: This paper proposes two algorithms: one for automatically computing a repair D' that satisfies a given set of CFDs, and the other for incrementally finding a repair in response to updates to a clean database.
Abstract: Two central criteria for data quality are consistency and accuracy. Inconsistencies and errors in a database often emerge as violations of integrity constraints. Given a dirty database D, one needs automated methods to make it consistent, i.e., find a repair D' that satisfies the constraints and "minimally" differs from D. Equally important is to ensure that the automatically-generated repair D' is accurate, or makes sense, i.e., D' differs from the "correct" data within a predefined bound. This paper studies effective methods for improving both data consistency and accuracy. We employ a class of conditional functional dependencies (CFDs) proposed in [6] to specify the consistency of the data, which are able to capture inconsistencies and errors beyond what their traditional counterparts can catch. To improve the consistency of the data, we propose two algorithms: one for automatically computing a repair D' that satisfies a given set of CFDs, and the other for incrementally finding a repair in response to updates to a clean database. We show that both problems are intractable. Although our algorithms are necessarily heuristic, we experimentally verify that the methods are effective and efficient. Moreover, we develop a statistical method that guarantees that the repairs found by the algorithms are accurate above a predefined rate without incurring excessive user interaction.

Journal ArticleDOI
TL;DR: The results show that the HPSO algorithm can effectively accelerate the convergence rate and can more quickly reach the optimum design than the two other algorithms.

Journal ArticleDOI
Yijun Sun1
TL;DR: This paper proposes an iterative RELIEF (I-RELIEF) algorithm to alleviate the deficiencies of RELIEf by exploring the framework of the expectation-maximization algorithm.
Abstract: RELIEF is considered one of the most successful algorithms for assessing the quality of features. In this paper, we propose a set of new feature weighting algorithms that perform significantly better than RELIEF, without introducing a large increase in computational complexity. Our work starts from a mathematical interpretation of the seemingly heuristic RELIEF algorithm as an online method solving a convex optimization problem with a margin-based objective function. This interpretation explains the success of RELIEF in real application and enables us to identify and address its following weaknesses. RELIEF makes an implicit assumption that the nearest neighbors found in the original feature space are the ones in the weighted space and RELIEF lacks a mechanism to deal with outlier data. We propose an iterative RELIEF (I-RELIEF) algorithm to alleviate the deficiencies of RELIEF by exploring the framework of the expectation-maximization algorithm. We extend I-RELIEF to multiclass settings by using a new multiclass margin definition. To reduce computational costs, an online learning algorithm is also developed. Convergence analysis of the proposed algorithms is presented. The results of large-scale experiments on the UCI and microarray data sets are reported, which demonstrate the effectiveness of the proposed algorithms, and verify the presented theoretical results

Journal ArticleDOI
TL;DR: A relaxation method is described which yields an easily computable upper bound on the optimal solution of portfolio selection, and a heuristic method for finding a suboptimal portfolio which is based on solving a small number of convex optimization problems.
Abstract: We consider the problem of portfolio selection, with transaction costs and constraints on exposure to risk. Linear transaction costs, bounds on the variance of the return, and bounds on different shortfall probabilities are efficiently handled by convex optimization methods. For such problems, the globally optimal portfolio can be computed very rapidly. Portfolio optimization problems with transaction costs that include a fixed fee, or discount breakpoints, cannot be directly solved by convex optimization. We describe a relaxation method which yields an easily computable upper bound via convex optimization. We also describe a heuristic method for finding a suboptimal portfolio, which is based on solving a small number of convex optimization problems (and hence can be done efficiently). Thus, we produce a suboptimal solution, and also an upper bound on the optimal solution. Numerical experiments suggest that for practical problems the gap between the two is small, even for large problems involving hundreds of assets. The same approach can be used for related problems, such as that of tracking an index with a portfolio consisting of a small number of assets.

Journal ArticleDOI
TL;DR: This work proposes an exact method for reducing the size of weighted (directed and undirected) complex networks while maintaining invariant its modularity, and compares the modularity obtained by using the Extremal Optimization algorithm, before and after the size reduction.
Abstract: The ubiquity of modular structure in real-world complex networks is being the focus of attention in many trials to understand the interplay between network topology and functionality. The best approaches to the identification of modular structure are based on the optimization of a quality function known as modularity. However this optimization is a hard task provided that the computational complexity of the problem is in the NP-hard class. Here we propose an exact method for reducing the size of weighted (directed and undirected) complex networks while maintaining invariant its modularity. This size reduction allows the heuristic algorithms that optimize modularity for a better exploration of the modularity landscape. We compare the modularity obtained in several real complex-networks by using the Extremal Optimization algorithm, before and after the size reduction, showing the improvement obtained. We speculate that the proposed analytical size reduction could be extended to an exact coarse graining of the network in the scope of real-space renormalization.

Journal ArticleDOI
TL;DR: A heuristic combining the adaptative memory principle, a tabu search method for the solution of subproblems, and integer programming is proposed, which indicates the validity of the technique to MDVRPI.

Journal ArticleDOI
TL;DR: An efficient variable neighborhood search heuristic for the capacitated vehicle routing problem to design least cost routes for a fleet of identically capacitated vehicles to service geographically scattered customers with known demands is presented.

Journal ArticleDOI
TL;DR: A heuristic method based on artificial neural networks (NN) is applied in order to trace out the efficient frontier associated to the portfolio selection problem, considering a generalization of the standard Markowitz mean-variance model which includes cardinality and bounding constraints.

Proceedings ArticleDOI
01 May 2007
TL;DR: A stochastic geometry model is presented for the performance analysis and the planning of dense IEEE 802.11 networks to take the effect of interferences and that of CSMA into account within this dense network context.
Abstract: This paper presents a stochastic geometry model for the performance analysis and the planning of dense IEEE 802.11 networks. This model allows one to propose heuristic formulas for various properties of such networks like the probability for users to be covered, the probability for access points to be granted access to the channel or the average long term throughput provided to end-users. The main merit of this model is to take the effect of interferences and that of CSMA into account within this dense network context. This analytic model, which is based on Matern point processes, is partly validated against simulation. It is then used to assess various properties of such networks. We show for instance how the long term throughput obtained by end-users behaves when the access point density increases. We also briefly show how to use this model for the planning of managed networks and for the economic modeling of unplanned networks.

Journal ArticleDOI
TL;DR: A cooperative metaheuristic to solve the location-routing problem with capacitated routes and depots is presented and it is shown that this meta heuristic outperforms other methods on various kinds of instances.
Abstract: Most of the time in a distribution system, depot location and vehicle routing are interdependent, and recent studies have shown that the overall system cost may be excessive if routing decisions are ignored when locating depots. The location-routing problem (LRP) overcomes this drawback by simultaneously tackling location and routing decisions. This paper presents a cooperative metaheuristic to solve the LRP with capacitated routes and depots. The principle is to alternate between a depot location phase and a routing phase, exchanging information on the most promising edges. In the first phase, the routes and their customers are aggregated into supercustomers, leading to a facility-location problem, which is then solved by a Lagrangean relaxation of the assignment constraints. In the second phase, the routes from the resulting multidepot vehicle-routing problem (VRP) are improved using a granular tabu search (GTS) heuristic. At the end of each global iteration, information about the edges most often used is recorded to be used in the following phases. The method is evaluated on three sets of randomly generated instances and compared with other heuristics and a lower bound. Solutions are obtained in a reasonable amount of time for such a strategic problem and show that this metaheuristic outperforms other methods on various kinds of instances.

Journal ArticleDOI
01 Feb 2007
TL;DR: A new memetic algorithm (MA) for multiobjective (MO) optimization is proposed, which combines the global search ability of particle swarm optimization with a synchronous local search heuristic for directed local fine-tuning.
Abstract: In this paper, a new memetic algorithm (MA) for multiobjective (MO) optimization is proposed, which combines the global search ability of particle swarm optimization with a synchronous local search heuristic for directed local fine-tuning. A new particle updating strategy is proposed based upon the concept of fuzzy global-best to deal with the problem of premature convergence and diversity maintenance within the swarm. The proposed features are examined to show their individual and combined effects in MO optimization. The comparative study shows the effectiveness of the proposed MA, which produces solution sets that are highly competitive in terms of convergence, diversity, and distribution

Journal ArticleDOI
TL;DR: Heuristic algorithms to solve the vehicle routing problem with simultaneous pick-up and delivery approximately in a small amount of computing time are presented and constructive algorithms, local search algorithms and tabu search algorithms are presented.

Book ChapterDOI
01 Jan 2007
TL;DR: This chapter presents a comprehensive overview of the available exact and heuristic algorithms for the VRP, most of which have been adapted to solve other variants.
Abstract: Publisher Summary This chapter discusses some of the most important vehicle routing problem types. The vehicle routing problem lies at the heart of distribution management. It is faced each day by thousands of companies and organizations engaged in the delivery and collection of goods or people. Because conditions vary from one setting to the next, the objectives and constraints encountered in practice are highly variable. Most algorithmic research and software development in this area focus on a limited number of prototype problems. By building enough flexibility in optimization systems, these can be adapted to various practical contexts. The classical vehicle routing problem (VRP) is one of the most popular problems in combinatorial optimization, and its study has given rise to several exact and heuristic solution techniques of general applicability. The chapter presents a comprehensive overview of the available exact and heuristic algorithms for the VRP, most of which have been adapted to solve other variants.

Journal Article
TL;DR: This paper reviews the major contributions to the Motion Planning field throughout a 35-year period, from classic approaches to heuristic algorithms, and concludes with comparative tables and graphs demonstrating the frequency of each MP method’s application.
Abstract: This paper reviews the major contributions to the Motion Planning (MP) field throughout a 35-year period, from classic approaches to heuristic algorithms Due to the NP-Hardness of the MP problem, heuristic methods have outperformed the classic approaches and have gained wide popularity After surveying around 1400 papers in the field, the amount of existing works for each method is identified and classified Especially, the history and applications of numerous heuristic methods in MP is investigated The paper concludes with comparative tables and graphs demonstrating the frequency of each MP method’s application, and so can be used as a guideline for MP researchers Keywords—Robot motion planning, Heuristic algorithms

Journal ArticleDOI
TL;DR: The main aim of this paper is to rigorously derive the updating formula of the k-modes clustering algorithm with the new dissimilarity measure and the convergence of the algorithm under the optimization framework.
Abstract: This correspondence describes extensions to the k-modes algorithm for clustering categorical data. By modifying a simple matching dissimilarity measure for categorical objects, a heuristic approach was developed in (Z. He, et al., 2005), (O. San, et al., 2004) which allows the use of the k-modes paradigm to obtain a cluster with strong intrasimilarity and to efficiently cluster large categorical data sets. The main aim of this paper is to rigorously derive the updating formula of the k-modes clustering algorithm with the new dissimilarity measure and the convergence of the algorithm under the optimization framework

Journal ArticleDOI
TL;DR: Integer programming and network flow-based lower-bounding methods that can solve moderately large instances of the WTA problem optimally and obtain almost optimal solutions of fairly large instances within a few seconds are suggested.
Abstract: The weapon-target assignment (WTA) problem is a fundamental problem arising in defense-related applications of operations research. This problem consists of optimally assigning n weapons to m targets so that the total expected survival value of the targets after all the engagements is minimal. The WTA problem can be formulated as a nonlinear integer programming problem and is known to be NP-complete. No exact methods exist for the WTA problem that can solve even small-size problems (for example, with 20 weapons and 20 targets). Although several heuristic methods have been proposed to solve the WTA problem, due to the absence of exact methods, no estimates are available on the quality of solutions produced by such heuristics. In this paper, we suggest integer programming and network flow-based lower-bounding methods that we obtain using a branch-and-bound algorithm for the WTA problem. We also propose a network flow-based construction heuristic and a very large-scale neighborhood (VLSN) search algorithm. We present computational results of our algorithms, which indicate that we can solve moderately large instances (up to 80 weapons and 80 targets) of the WTA problem optimally and obtain almost optimal solutions of fairly large instances (up to 200 weapons and 200 targets) within a few seconds.

Proceedings Article
22 Jul 2007
TL;DR: In this paper, a local search approach for algorithm configuration is presented, which can be used for minimising run-time in decision problems or for maximising solution quality in optimisation problems, with no limitation on the number of parameters.
Abstract: The determination of appropriate values for free algorithm parameters is a challenging and tedious task in the design of effective algorithms for hard problems. Such parameters include categorical choices (e.g., neighborhood structure in local search or variable/value ordering heuristics in tree search), as well as numerical parameters (e.g., noise or restart timing). In practice, tuning of these parameters is largely carried out manually by applying rules of thumb and crude heuristics, while more principled approaches are only rarely used. In this paper, we present a local search approach for algorithm configuration and prove its convergence to the globally optimal parameter configuration. Our approach is very versatile: it can, e.g., be used for minimising run-time in decision problems or for maximising solution quality in optimisation problems. It further applies to arbitrary algorithms, including heuristic tree search and local search algorithms, with no limitation on the number of parameters. Experiments in four algorithm configuration scenarios demonstrate that our automatically determined parameter settings always outperform the algorithm defaults, sometimes by several orders of magnitude. Our approach also shows better performance and greater flexibility than the recent CALIBRA system. Our ParamILS code, along with instructions on how to use it for tuning your own algorithms, is available on-line at http://www.cs.ubc.ca/labs/beta/Projects/ParamILS.

01 Jan 2007
TL;DR: In this article, the authors proposed a method to approximate the nonlinear objective function of the problem by means of piecewise-linear functions, so that UC can be approximated by an mixed-integer linear program (MILP).
Abstract: The short-term unit commitment (UC) problem in hydrothermal power generation is a large-scale, mixed-integer nonlinear program, which is difficult to solve efficiently, especially for large-scale instances. It is possible to approximate the nonlinear objective function of the problem by means of piecewise-linear functions, so that UC can be approximated by an mixed-integer linear program (MILP); applying the available efficient general-purpose MILP solvers to the resulting formulations, good quality solutions can be obtained in a relatively short amount of time. We build on this approach, presenting a novel way to approximating the nonlinear objective function based on a recently developed class of valid inequalities for the problem, called ldquoperspective cuts.rdquo At least for many realistic instances of a general basic formulation of UC, an MILP-based heuristic obtains comparable or slightly better solutions in less time when employing the new approach rather than the standard piecewise linearizations, while being not more difficult to implement and use. Furthermore, ldquodynamicrdquo formulations, whereby the approximation is iteratively improved, provide even better results if the approximation is appropriately controlled.

Book ChapterDOI
Eyal Even-Dar1, Asaf Shapira2
12 Dec 2007
TL;DR: A very simple and efficient algorithms are provided for solving the spread maximization problem in the context of the well studied probabilistic voter model and it is shown that the most natural heuristic solution, which picks the nodes in the network with the highest degree is indeed the optimal solution.
Abstract: We consider the spread maximization problem that was defined by Domingos and Richardson [6,15] In this problem, we are given a social network represented as a graph and are required to find the set of the most "influential" individuals that by introducing them with a new technology, we maximize the expected number of individuals in the network, later in time, that adopt the new technology This problem has applications in viral marketing, where a company may wish to spread the rumor of a new product via the most influential individuals in popular social networks such as Myspace and Blogsphere The spread maximization problem was recently studied in several models of social networks [10,11,13] In this short paper we study this problem in the context of the well studied probabilistic voter model We provide very simple and efficient algorithms for solving this problem An interesting special case of our result is that the most natural heuristic solution, which picks the nodes in the network with the highest degree, is indeed the optimal solution

Journal ArticleDOI
01 Feb 2007
TL;DR: The proposed method is based on a parallel implementation of a multipopulation steady-state GA involving local search heuristics that uses a variant of the maximal preservative crossover and the double-bridge move mutation to find high-quality solutions for the traveling salesman problem.
Abstract: This correspondence describes a hybrid genetic algorithm (GA) to find high-quality solutions for the traveling salesman problem (TSP). The proposed method is based on a parallel implementation of a multipopulation steady-state GA involving local search heuristics. It uses a variant of the maximal preservative crossover and the double-bridge move mutation. An effective implementation of the Lin-Kernighan heuristic (LK) is incorporated into the method to compensate for the GA's lack of local search ability. The method is validated by comparing it with the LK-Helsgaun method (LKH), which is one of the most effective methods for the TSP. Experimental results with benchmarks having up to 316 228 cities show that the proposed method works more effectively and efficiently than LKH when solving large-scale problems. Finally, the method is used together with the implementation of the iterated LK to find a new best tour (as of June 2, 2003) for a 1 904 711-city TSP challenge