scispace - formally typeset

Topic

Parameterized complexity

About: Parameterized complexity is a(n) research topic. Over the lifetime, 6681 publication(s) have been published within this topic receiving 133575 citation(s).


Papers
More filters
Journal ArticleDOI
TL;DR: Several properties of the graph-theoretic complexity are proved which show, for example, that complexity is independent of physical size and complexity depends only on the decision structure of a program.
Abstract: This paper describes a graph-theoretic complexity measure and illustrates how it can be used to manage and control program complexity. The paper first explains how the graph-theory concepts apply and gives an intuitive explanation of the graph concepts in programming terms. The control graphs of several actual Fortran programs are then presented to illustrate the correlation between intuitive complexity and the graph-theoretic complexity. Several properties of the graph-theoretic complexity are then proved which show, for example, that complexity is independent of physical size (adding or subtracting functional statements leaves complexity unchanged) and complexity depends only on the decision structure of a program.

4,749 citations

Book
06 Nov 1998
TL;DR: An approach to complexity theory which offers a means of analysing algorithms in terms of their tractability, and introduces readers to new classes of algorithms which may be analysed more precisely than was the case until now.
Abstract: An approach to complexity theory which offers a means of analysing algorithms in terms of their tractability. The authors consider the problem in terms of parameterized languages and taking "k-slices" of the language, thus introducing readers to new classes of algorithms which may be analysed more precisely than was the case until now. The book is as self-contained as possible and includes a great deal of background material. As a result, computer scientists, mathematicians, and graduate students interested in the design and analysis of algorithms will find much of interest.

3,578 citations

Book
01 Jan 2010
TL;DR: Fixed-Parameter Tractability.
Abstract: Fixed-Parameter Tractability.- Reductions and Parameterized Intractability.- The Class W[P].- Logic and Complexity.- Two Fundamental Hierarchies.- The First Level of the Hierarchies.- The W-Hierarchy.- The A- Hierarchy.- Kernelization and Linear Programming Techniques.- The Automata-Theoretic Approach.- Tree Width.- Planarity and Bounded Local Tree Width.- Homomorphisms and Embeddings.- Parameterized Counting Problems.- Bounded Fixed-Parameter Tractability.- Subexponential Fixed-Parameter Tractability.- Appendix, Background from Complexity Theory.- References.- Notation.- Index.

2,278 citations

Journal ArticleDOI
Abstract: J. F. Benders devised a clever approach for exploiting the structure of mathematical programming problems withcomplicating variables (variables which, when temporarily fixed, render the remaining optimization problem considerably more tractable). For the class of problems specifically considered by Benders, fixing the values of the complicating variables reduces the given problem to an ordinary linear program, parameterized, of course, by the value of the complicating variables vector. The algorithm he proposed for finding the optimal value of this vector employs a cutting-plane approach for building up adequate representations of (i) the extremal value of the linear program as a function of the parameterizing vector and (ii) the set of values of the parameterizing vector for which the linear program is feasible. Linear programming duality theory was employed to derive the natural families ofcuts characterizing these representations, and the parameterized linear program itself is used to generate what are usuallydeepest cuts for building up the representations.

2,000 citations

Journal ArticleDOI
Abstract: of the number of bits required to write down the observed data, has been reformulated to extend the classical maximum likelihood principle. The principle permits estimation of the number of the parameters in statistical models in addition to their values and even of the way the parameters appear in the models; i.e., of the model structures. The principle rests on a new way to interpret and construct a universal prior distribution for the integers, which makes sense even when the parameter is an individual object. Truncated realvalued parameters are converted to integers by dividing them by their precision, and their prior is determined from the universal prior for the integers by optimizing the precision. 1. Introduction. In this paper we study estimation based upon the principle of minimizing the total number of binary digits required to rewrite the observed data, when each observation is given with some precision. Instead of attempting at an absolutely shortest description, which would be futile, we look for the optimum relative to a class of parametrically given distributions. This Minimum Description Length (MDL) principle, which we introduced in a less comprehensive form in [25], turns out to degenerate to the more familiar Maximum Likelihood (ML) principle in case the number of parameters in the models is fixed, so that the description length of the parameters themselves can be ignored. In another extreme case, where the parameters determine the data, it similarly degenerates to Jaynes's principle of maximum entropy, [14]. But the main power of the new criterion is that it permits estimates of the entire model, its parameters, their number, and even the way the parameters appear in the model; i.e., the model structure. Hence, there will be no need to supplement the estimated parameters with a separate hypothesis test to decide whether a model is adequately parameterized or, perhaps, over parameterized.

1,701 citations


Network Information
Related Topics (5)
Robustness (computer science)

94.7K papers, 1.6M citations

86% related
Optimization problem

96.4K papers, 2.1M citations

85% related
Matrix (mathematics)

105.5K papers, 1.9M citations

84% related
Nonlinear system

208.1K papers, 4M citations

82% related
Cluster analysis

146.5K papers, 2.9M citations

82% related
Performance
Metrics
No. of papers in the topic in previous years
YearPapers
20228
2021414
2020557
2019474
2018463
2017451