scispace - formally typeset
Search or ask a question
Book ChapterDOI

Decoupling Network Optimization by Swarm Intelligence

TL;DR: In this chapter, the problem of decoupling network optimization is discussed in detail and Swarm intelligence is used for maintaining power integrity in high-speed systems.
Abstract: In this chapter, the problem of decoupling network optimization is discussed in detail. Swarm intelligence is used for maintaining power integrity in high-speed systems. The optimum number of capacitors and their values are selected to meet the target impedance of the system.
References
More filters
Proceedings ArticleDOI
01 Dec 2009
TL;DR: A new meta-heuristic algorithm, called Cuckoo Search (CS), is formulated, based on the obligate brood parasitic behaviour of some cuckoo species in combination with the Lévy flight behaviour ofSome birds and fruit flies, for solving optimization problems.
Abstract: In this paper, we intend to formulate a new meta-heuristic algorithm, called Cuckoo Search (CS), for solving optimization problems. This algorithm is based on the obligate brood parasitic behaviour of some cuckoo species in combination with the Levy flight behaviour of some birds and fruit flies. We validate the proposed algorithm against test functions and then compare its performance with those of genetic algorithms and particle swarm optimization. Finally, we discuss the implication of the results and suggestion for further research.

5,521 citations

Book
01 Feb 2008
TL;DR: This book reviews and introduces the state-of-the-art nature-inspired metaheuristic algorithms in optimization, including genetic algorithms, bee algorithms, particle swarm optimization, simulated annealing, ant colony optimization, harmony search, and firefly algorithms.
Abstract: Modern metaheuristic algorithms such as bee algorithms and harmony search start to demonstrate their power in dealing with tough optimization problems and even NP-hard problems. This book reviews and introduces the state-of-the-art nature-inspired metaheuristic algorithms in optimization, including genetic algorithms, bee algorithms, particle swarm optimization, simulated annealing, ant colony optimization, harmony search, and firefly algorithms. We also briefly introduce the photosynthetic algorithm, the enzyme algorithm, and Tabu search. Worked examples with implementation have been used to show how each algorithm works. This book is thus an ideal textbook for an undergraduate and/or graduate course. As some of the algorithms such as the harmony search and firefly algorithms are at the forefront of current research, this book can also serve as a reference book for researchers.

3,626 citations

Book ChapterDOI
26 Oct 2009
TL;DR: In this article, a new Firefly Algorithm (FA) was proposed for multimodal optimization applications. And the proposed FA was compared with other metaheuristic algorithms such as particle swarm optimization (PSO).
Abstract: Nature-inspired algorithms are among the most powerful algorithms for optimization. This paper intends to provide a detailed description of a new Firefly Algorithm (FA) for multimodal optimization applications. We will compare the proposed firefly algorithm with other metaheuristic algorithms such as particle swarm optimization (PSO). Simulations and results indicate that the proposed firefly algorithm is superior to existing metaheuristic algorithms. Finally we will discuss its applications and implications for further research.

3,436 citations

Book
01 Jan 1987
TL;DR: This second edition has been updated and expanded to cover recent developments in applications and theory, including an elegant NP completeness argument by Uwe Naumann and a brief introduction to scarcity, a generalization of sparsity.
Abstract: Algorithmic, or automatic, differentiation (AD) is a growing area of theoretical research and software development concerned with the accurate and efficient evaluation of derivatives for function evaluations given as computer programs. The resulting derivative values are useful for all scientific computations that are based on linear, quadratic, or higher order approximations to nonlinear scalar or vector functions. AD has been applied in particular to optimization, parameter identification, nonlinear equation solving, the numerical integration of differential equations, and combinations of these. Apart from quantifying sensitivities numerically, AD also yields structural dependence information, such as the sparsity pattern and generic rank of Jacobian matrices. The field opens up an exciting opportunity to develop new algorithms that reflect the true cost of accurate derivatives and to use them for improvements in speed and reliability. This second edition has been updated and expanded to cover recent developments in applications and theory, including an elegant NP completeness argument by Uwe Naumann and a brief introduction to scarcity, a generalization of sparsity. There is also added material on checkpointing and iterative differentiation. To improve readability the more detailed analysis of memory and complexity bounds has been relegated to separate, optional chapters.The book consists of three parts: a stand-alone introduction to the fundamentals of AD and its software; a thorough treatment of methods for sparse problems; and final chapters on program-reversal schedules, higher derivatives, nonsmooth problems and iterative processes. Each of the 15 chapters concludes with examples and exercises. Audience: This volume will be valuable to designers of algorithms and software for nonlinear computational problems. Current numerical software users should gain the insight necessary to choose and deploy existing AD software tools to the best advantage. Contents: Rules; Preface; Prologue; Mathematical Symbols; Chapter 1: Introduction; Chapter 2: A Framework for Evaluating Functions; Chapter 3: Fundamentals of Forward and Reverse; Chapter 4: Memory Issues and Complexity Bounds; Chapter 5: Repeating and Extending Reverse; Chapter 6: Implementation and Software; Chapter 7: Sparse Forward and Reverse; Chapter 8: Exploiting Sparsity by Compression; Chapter 9: Going beyond Forward and Reverse; Chapter 10: Jacobian and Hessian Accumulation; Chapter 11: Observations on Efficiency; Chapter 12: Reversal Schedules and Checkpointing; Chapter 13: Taylor and Tensor Coefficients; Chapter 14: Differentiation without Differentiability; Chapter 15: Implicit and Iterative Differentiation; Epilogue; List of Figures; List of Tables; Assumptions and Definitions; Propositions, Corollaries, and Lemmas; Bibliography; Index

2,920 citations

Journal ArticleDOI
TL;DR: This paper presents a more extensive comparison study using some standard test functions and newly designed stochastic test functions to apply the CS algorithm to solve engineering design optimisation problems, including the design of springs and welded beam structures.
Abstract: A new metaheuristic optimisation algorithm, called cuckoo search (CS), was developed recently by Yang and Deb (2009). This paper presents a more extensive comparison study using some standard test functions and newly designed stochastic test functions. We then apply the CS algorithm to solve engineering design optimisation problems, including the design of springs and welded beam structures. The optimal solutions obtained by CS are far better than the best solutions obtained by an efficient particle swarm optimiser. We will discuss the unique search features used in CS and the implications for further research.

1,339 citations