scispace - formally typeset
Search or ask a question

Showing papers on "Time complexity published in 1992"


Book ChapterDOI
07 Jul 1992
TL;DR: Comparison with other feature selection algorithms shows Relief's advantages in terms of learning time and the accuracy of the learned concept, suggesting Relief's practicality.
Abstract: In real-world concept learning problems, the representation of data often uses many features, only a few of which may be related to the target concept. In this situation, feature selection is important both to speed up learning and to improve concept quality. A new feature selection algorithm Relief uses a statistical method and avoids heuristic search. Relief requires linear time in the number of given features and the number of training instances regardless of the target concept to be learned. Although the algorithm does not necessarily find the smallest subset of features, the size tends to be small because only statistically relevant features are selected. This paper focuses on empirical test results in two artificial domains; the LED Display domain and the Parity domain with and without noise. Comparison with other feature selection algorithms shows Relief's advantages in terms of learning time and the accuracy of the learned concept, suggesting Relief's practicality.

2,908 citations


Proceedings Article
12 Jul 1992
TL;DR: A new algorithm Rellef is introduced which selects relevant features using a statistical method and is accurate even if features interact, and is noise-tolerant, suggesting a practical approach to feature selection for real-world problems.
Abstract: For real-world concept learning problems, feature selection is important to speed up learning and to improve concept quality We review and analyze past approaches to feature selection and note their strengths and weaknesses We then introduce and theoretically examine a new algorithm Rellef which selects relevant features using a statistical method Relief does not depend on heuristics, is accurate even if features interact, and is noise-tolerant It requires only linear time in the number of given features and the number of training instances, regardless of the target concept complexity The algorithm also has certain limitations such as nonoptimal feature set size Ways to overcome the limitations are suggested We also report the test results of comparison between Relief and other feature selection algorithms The empirical results support the theoretical analysis, suggesting a practical approach to feature selection for real-world problems

1,910 citations


Proceedings ArticleDOI
24 Oct 1992
TL;DR: Agarwal et al. as discussed by the authors showed that the MAXSNP-hard problem does not have polynomial-time approximation schemes unless P=NP, and for some epsilon > 0 the size of the maximal clique in a graph cannot be approximated within a factor of n/sup 1/ε / unless P = NP.
Abstract: The class PCP(f(n),g(n)) consists of all languages L for which there exists a polynomial-time probabilistic oracle machine that used O(f(n)) random bits, queries O(g(n)) bits of its oracle and behaves as follows: If x in L then there exists an oracle y such that the machine accepts for all random choices but if x not in L then for every oracle y the machine rejects with high probability. Arora and Safra (1992) characterized NP as PCP(log n, (loglogn)/sup O(1)/). The authors improve on their result by showing that NP=PCP(logn, 1). The result has the following consequences: (1) MAXSNP-hard problems (e.g. metric TSP, MAX-SAT, MAX-CUT) do not have polynomial time approximation schemes unless P=NP; and (2) for some epsilon >0 the size of the maximal clique in a graph cannot be approximated within a factor of n/sup epsilon / unless P=NP. >

1,277 citations


Journal ArticleDOI
TL;DR: It is proven that when both randomization and interaction are allowed, the proofs that can be verified in polynomial time are exactly those proofs that could be generated with polynometric space.
Abstract: In this paper, it is proven that when both randomization and interaction are allowed, the proofs that can be verified in polynomial time are exactly those proofs that can be generated with polynomial space.

821 citations


Journal ArticleDOI
06 Jan 1992
TL;DR: Two string distance functions that are computable in linear time give a lower bound for the edit distance (in the unit cost model), which leads to fast hybrid algorithms for the edited distance based string matching.
Abstract: We study approximate string matching in connection with two string distance functions that are computable in linear time. The first function is based on the so-called $q$-grams. An algorithm is given for the associated string matching problem that finds the locally best approximate occurences of pattern $P$, $|P|=m$, in text $T$, $|T|=n$, in time $O(n\log (m-q))$. The occurences with distance $\leq k$ can be found in time $O(n\log k)$. The other distance function is based on finding maximal common substrings and allows a form of approximate string matching in time $O(n)$. Both distances give a lower bound for the edit distance (in the unit cost model), which leads to fast hybrid algorithms for the edit distance based string matching.

665 citations


Journal ArticleDOI
TL;DR: It is shown that the class of languages having tow-prover interactive proof systems is nondeterministic exponential time and that to prove membership in languages inEXP, the honest provers need the power ofEXP only.
Abstract: We determine the exact power of two-prover interactive proof systems introduced by Ben-Or, Goldwasser, Kilian, and Wigderson (1988). In this system, two all-powerful noncommunicating provers convince a randomizing polynomial time verifier in polynomial time that the inputx belongs to the languageL. We show that the class of languages having tow-prover interactive proof systems is nondeterministic exponential time. We also show that to prove membership in languages inEXP, the honest provers need the power ofEXP only. The first part of the proof of the main result extends recent techniques of polynomial extrapolation used in the single prover case by Lund, Fortnow, Karloff, Nisan, and Shamir. The second part is averification scheme for multilinearity of a function in several variables held by an oracle and can be viewed as an independent result onprogram verification. Its proof rests on combinatorial techniques employing a simple isoperimetric inequality for certain graphs:

601 citations


Journal ArticleDOI
TL;DR: Pseudorandom generators are constructed which convertO(SlogR) truly random bits toR bits that appear random to any algorithm that runs inSPACE(S) that can be simulated using onlyO(Slogn) random bits.
Abstract: Pseudorandom generators are constructed which convertO(SlogR) truly random bits toR bits that appear random to any algorithm that runs inSPACE(S). In particular, any randomized polynomial time algorithm that runs in spaceS can be simulated using onlyO(Slogn) random bits. An application of these generators is an explicit construction of universal traversal sequences (for arbitrary graphs) of lengthn O(logn). The generators constructed are technically stronger than just appearing random to spacebounded machines, and have several other applications. In particular, applications are given for “deterministic amplification” (i.e. reducing the probability of error of randomized algorithms), as well as generalizations of it.

538 citations


Journal ArticleDOI
TL;DR: It is shown that any deterministic broadcast protocol requires 8(n) time-slots, even if the network has diameter 3, and n is known to all processors, which demonstrates an exponential gap in complexity between randomization and determinism.

492 citations


Journal ArticleDOI
TL;DR: An algorithm to solve the economic lot sizing problem in O(n log n) time is presented and it is shown how the Wagner-Whitin case can even be solved in linear time.
Abstract: We consider the n-period economic lot sizing problem, where the cost coefficients are not restricted in sign. In their seminal paper, H. M. Wagner and T. M. Whitin proposed an O(n2) algorithm for the special case of this problem, where the marginal production costs are equal in all periods and the unit holding costs are nonnegative. It is well known that their approach can also be used to solve the general problem, without affecting the complexity of the algorithm. In this paper, we present an algorithm to solve the economic lot sizing problem in O(n log n) time, and we show how the Wagner-Whitin case can even be solved in linear time. Our algorithm can easily be explained by a geometrical interpretation and the time bounds are obtained without the use of any complicated data structure. Furthermore, we show how Wagner and Whitin's and our algorithm are related to algorithms that solve the dual of the simple plant location formulation of the economic lot sizing problem.

490 citations


Journal ArticleDOI
TL;DR: Efficient algorithms for computing approximate traveling salesman tours in multidimensional point sets and three local optimizations are described, which are able to solve uniform planar million-city traveling salesman problems to within a few percent of optimal in several midicomputer CPU hours.
Abstract: This paper describes efficient algorithms for computing approximate traveling salesman tours in multidimensional point sets. We describe implementations of a dozen starting heuristics (including Nearest Neighbor and Farthest Insertion) and three local optimizations (including 2-Opt and 3-Opt). Experiments indicate that most of the algorithms run in O(N log N) time on uniform data sets, and many run almost as fast on very nonuniform data. The program that implements the algorithms is able to solve uniform planar million-city traveling salesman problems to within a few percent of optimal in several midicomputer CPU hours. The algorithms and program apply to many distance metrics and dimensions. INFORMS Journal on Computing, ISSN 1091-9856, was published as ORSA Journal on Computing from 1989 to 1995 under ISSN 0899-1499.

486 citations


Proceedings ArticleDOI
24 Oct 1992
TL;DR: The authors give a new characterization of NP: the class NP contains exactly those languages L for which membership proofs can be verified probabilistically in polynomial time using logarithmic number of random bits and sub-logarital number of queries to the proof.
Abstract: The authors give a new characterization of NP: the class NP contains exactly those languages L for which membership proofs (a proof that an input x is in L) can be verified probabilistically in polynomial time using logarithmic number of random bits and sub-logarithmic number of queries to the proof. This is a non-relativizing characterization of NP. They discuss implications of this characterization; specifically, they show that approximating clique (or independent set) is NP-hard. >

Journal ArticleDOI
TL;DR: A recursion-theoretic characterization of FP which describes polynomial time computation independently of any externally imposed resource bounds, and avoids the explicit size bounds on recursion of Cobham.
Abstract: We give a recursion-theoretic characterization of FP which describes polynomial time computation independently of any externally imposed resource bounds. In particular, this syntactic characterization avoids the explicit size bounds on recursion (and the initial function 2|x|·|y|) of Cobham.

Journal ArticleDOI
TL;DR: An algorithm for computing the edge-connectivity of a multigraph, where p ( \leqq | E | )$ is the number of pairs of nodes between which G has an edge, is proposed and consists only of times of graph searches and edge contractions.
Abstract: Given an undirected graph $G = ( V,E )$, it is known that its edge-connectivity $\lambda ( G )$ can be computed by solving $O( | V | )$ max-flow problems The best time bounds known for the problem

Journal ArticleDOI
TL;DR: It is shown that anyk-connected graphG = (V, E) has a sparsek- connected spanning subgraphG′ = ( V, E′) with ¦E′¦ =O(k¦V¦) by presenting anO(¦E¦)-time algorithm to find one such subgraph, where connectivity stands for either edge-connectivity or node-connectivities.
Abstract: We show that anyk-connected graphG = (V, E) has a sparsek-connected spanning subgraphG′ = (V, E′) with ¦E′¦ =O(k¦V¦) by presenting anO(¦E¦)-time algorithm to find one such subgraph, where connectivity stands for either edge-connectivity or node-connectivity. By using this algorithm as preprocessing, the time complexities of some graph problems related to connectivity can be improved. For example, the current best time boundO(max{k 2¦V¦1/2,k¦V¦}¦E¦) to determine whether node-connectivityK(G) of a graphG = (V, E) is larger than a given integerk or not can be reduced toO(max{k 3¦V¦3/2,k 2¦V¦2}).

Journal ArticleDOI
TL;DR: To analyze the complexity of the algorithm, an amortization argument based on a new combinatorial theorem on line arrangements is used.
Abstract: The main contribution of this work is an O(n log n + k)-time algorithm for computing all k intersections among n line segments in the plane. This time complexity is easily shown to be optimal. Within the same asymptotic cost, our algorithm can also construct the subdivision of the plane defined by the segments and compute which segment (if any) lies right above (or below) each intersection and each endpoint. The algorithm has been implemented and performs very well. The storage requirement is on the order of n + k in the worst case, but it is considerably lower in practice. To analyze the complexity of the algorithm, an amortization argument based on a new combinatorial theorem on line arrangements is used.

Journal ArticleDOI
TL;DR: In this paper, the authors prove lower bounds of the form Ω(n · c−k), for the number of bits that need to be exchanged in order to compute some (explicitly given) polynomial time computable functions.

Journal ArticleDOI
TL;DR: It is proved that any functional term of appropriate type actually encodes a polynomial-time algorithm and that conversely any polynometric-time function can be obtained in this way.

Proceedings ArticleDOI
04 May 1992
TL;DR: The results establish that strong typing is crucial to achieving a useful demarcation between decidable and undecidable safety, and ternary monotonic commands are critical for tractable safety analysis.
Abstract: The typed access matrix (TAM) model is defined by introducing the notion of strong typing into the Harrison, Ruzzo, and Ullman model (HRU) (M. H. Harrison et al., 1978). It is shown that monotonic TAM (MTAM) has decidable, but NP-hard, safety for its acyclic creation cases. It is further shown that ternary MTAM has polynomial time safety analysis for its acyclic cases, even though it is, in general, equivalent to MTAM. Ternary MTAM thus has strong safety properties. The expressive power of ternary MTAM has been shown to be equivalent to MTAM in general. The results establish that strong typing is crucial to achieving a useful demarcation between decidable and undecidable safety, and ternary monotonic commands are critical for tractable safety analysis. >

Journal ArticleDOI
TL;DR: A predicate calculus in which graph problems can be expressed and any problem possessing such an expression can be solved in linear time on any recursively constructed graph, once its decomposition tree is known.
Abstract: This paper describes a predicate calculus in which graph problems can be expressed. Any problem possessing such an expression can be solved in linear time on any recursively constructed graph, once its decomposition tree is known. Moreover, the linear-time algorithm can be generatedautomatically from the expression, because all our theorems are proved constructively. The calculus is founded upon a short list of particularly primitive predicates, which in turn are combined by fundamental logical operations. This framework is rich enough to include the vast majority of known linear-time solvable problems. We have obtained these results independently of similar results by Courcelle [11], [12], through utilization of the framework of Bernet al. [6]. We believe our formalism is more practical for programmers who would implement the automatic generation machinery, and more readily understood by many theorists.

Journal ArticleDOI
TL;DR: Two modifications of the Lagarias-Odlyzko algorithm, either one of which would solve almost all problems of density<0.9408 ... if it could find shortest non-zero vectors in lattices if they were combined with known lattice basis reduction algorithms.
Abstract: The general subset sum problem is NP-complete. However, there are two algorithms, one due to Brickell and the other to Lagarias and Odlyzko, which in polynomial time solve almost all subset sum problems of sufficiently low density. Both methods rely on basis reduction algorithms to find short non-zero vectors in special lattices. The Lagarias-Odlyzko algorithm would solve almost all subset sum problems of density<0.6463 ... in polynomial time if it could invoke a polynomial-time algorithm for finding the shortest non-zero vector in a lattice. This paper presents two modifications of that algorithm, either one of which would solve almost all problems of density<0.9408 ... if it could find shortest non-zero vectors in lattices. These modifications also yield dramatic improvements in practice when they are combined with known lattice basis reduction algorithms.

Journal ArticleDOI
01 May 1992-Networks
TL;DR: Algorithms are derived for the case where, for any time t, the number of jobs that can be executed at that time is bounded, and for the special cases where all of the processing times are 0, all ofThe release times ri are0, and all ofthe deadlines di are infinite.
Abstract: Consider a complete directed graph in which each arc has a given length. There is a set ofjobs, each job i located at some node of the graph, with an associated processing time hi, and whose execution has to start within a prespecified time window [r;, di]. We have a single server that can move on the arcs of the graph, at unit speed, and that has to execute all of the jobs within their respective time windows. We consider the following two problems: (a) minimize the time by which all jobs are executed (traveling salesman problem) and (b) minimize the sum of the waiting times of the jobs (traveling repairman problem). We focus on the following two special cases: (a) The jobs are located on a line and (b) the number of nodes of the graph is bounded by some integer constant B. Furthermore, we consider in detail the special cases where (a) all of the processing times are 0, (b) all of the release times ri are 0, and (c) all of the deadlines di are infinite. For many of the resulting problem combinations, we settle their complexity either by establishing NP-completeness or by presenting polynomial (or pseudopolynomial) time algorithms. Finally, we derive algorithms for the case where, for any time t, the number of jobs that can be executed at that time is bounded.

Book ChapterDOI
13 Jul 1992
TL;DR: It is shown that the general problem of computation of phytogenies for species sets is NP-Complete, and that the various finite-state approaches for bounded treewidth cannot be applied to the fixed-parameter forms of the problem.
Abstract: One of the major efforts in molecular biology is the computation of phytogenies for species sets. A longstanding open problem in this area is called the Perfect Phylogeny problem. For almost two decades the complexity of this problem remained open, with progress limited to polynomial time algorithms for a few special cases, and many relaxations of the problem shown to be NP-Complete. From an applications point of view, the problem is of interest both in its general form, where the number of characters may vary, and in its fixed-parameter form. The Perfect Phylogeny problem has been shown to be equivalent to the problem of triangulating colored graphs[30]. It has also been shown recently that for a given fixed number of characters the yes-instances have bounded treewidth[45], opening the possibility of applying methodologies for bounded treewidth to the fixed-parameter form of the problem. We show that the Perfect Phylogeny problem is difficult in two different ways. We show that the general problem is NP-Complete, and we show that the various finite-state approaches for bounded treewidth cannot be applied to the fixed-parameter forms of the problem.

Proceedings ArticleDOI
24 Oct 1992
TL;DR: The authors provide data structures that maintain a graph as edges are inserted and deleted, and keep track of the following properties: minimum spanning forests, best swap, graph connectivity, and graph
Abstract: The authors provide data structures that maintain a graph as edges are inserted and deleted, and keep track of the following properties: minimum spanning forests, best swap, graph connectivity, and graph 2-edge-connectivity, in time O(n/sup 1/2/log(m/n)) per change; 3-edge-connectivity, in time O(n/sup 2/3/) per change; 4-edge-connectivity, in time O(n alpha (n)) per change; k-edge-connectivity, in time O(n log n) per change; bipartiteness, 2-vertex-connectivity, and 3-vertex-connectivity, in time O(n log(m/n)) per change; and 4-vertex-connectivity, in time O(n log(m/n)+n alpha (n)) per change. Further results speed up the insertion times to match the bounds of known partially dynamic algorithms. The algorithms are based on a technique that transforms algorithms for sparse graphs into ones that work on any graph, which they call sparsification. >

Proceedings ArticleDOI
Bruce Reed1
01 Jul 1992
TL;DR: This approximate separator algorithm can be used to develop and develop and O(n) algorithm for determining if G has a tree decomposition of width at most k (for fixed k) and finding such a Tree decomposition if it exists.
Abstract: We show that for any fixed k, there is a linear-time algorithm which given a graph G either: (i) finds a cutset X of G with |X| ≤ k such that no component of G–X contains more than 3/4|G–X| vertices, or (ii) determines that for any set X of vertices of G with |X| ≤ k, there is a component of G–X which contains more than 2/3|G–X| vertices.This approximate separator algorithm can be used to develop and O(n log n algorithm for determining if G has a tree decomposition of width at most k (for fixed k) and finding such a tree decomposition if it exists.

Journal ArticleDOI
TL;DR: The structural result is extended and used to solve a general problem of which the Frobenius problem is a special case, and a polynomial time algorithm is developed for finding the covering radius of any polytope in a fixed number of dimensions.
Abstract: This paper considers the “Frobenius problem”: Givenn natural numbersa1,a2,...an such that their greatest common divisor is 1, find the largest natural number that is not expressible as a nonnegative integer combination of them. This problem can be seen to be NP-hard. For the casesn=2,3 polynomial time algorithms, are known to solve it. Here a polynomial time algorithm is given for every fixedn. This is done by first proving an exact relation between the Frobenius problem and a geometric concept called the “covering radius”. Then a polynomial time algorithm is developed for finding the covering radius of any polytope in a fixed number of dimensions. The last algorithm relies on a structural theorem proved here that describes for any polytopeK, the setK+ℤh={x∶x∈ℝn;x=y+z;y∈K;z∈ℤn} which is the portion of space covered by all lattice translates ofK. The proof of the structural theorem relies on some recent developments in the Geometry of Numbers. In particular, it uses a theorem of Kannan and Lovasz [11], bounding the width of lattice-point-free convex bodies and the techniques of Kannan, Lovasz and Scarf [12] to study the shapes of a polyhedron obtained by translating each facet parallel, to itself. The concepts involved are defined from first principles. In a companion paper [10], I extend the structural result and use that to solve a general problem of which the Frobenius problem is a special case.

Journal ArticleDOI
TL;DR: The complexity of these problems when G is restricted to be a partial k -tree is discussed, and a polynomial time algorithm is given for the n disjoint connecting paths problem restricted topartial k -trees (with n part of input).

Journal ArticleDOI
TL;DR: The scope of the theory of average case complexity is widened to other basic questions in computational complexity, and definitions and basic theorems regarding other complexity classes such as average log-space are provided.

Journal ArticleDOI
TL;DR: Dynamic programming solutions to a number of different recurrence equations for sequence comparison and for RNA secondary structure prediction are considered, when the weight functions used in the recurrences are taken to be linear.
Abstract: Dynamic programming solutions to a number of different recurrence equations for sequence comparison and for RNA secondary structure prediction are considered. These recurrences are defined over a number of points that is quadratic in the input size; however only a sparse set matters for the result. Efficient algorithms for these problems are given, when the weight functions used in the recurrences are taken to be linear. The time complexity of the algorithms depends almost linearly on the number of points that need to be considered; when the problems are sparse this results in a substantial speed-up over known algorithms.

DissertationDOI
01 Jan 1992
TL;DR: It is shown how the BH algorithm can be adapted to execute in parallel, and the performance of the parallel version of the algorithm is analyzed, finding that the overhead is due primarily to interprocessor synchronization delays and redundant computation.
Abstract: Recent algorithmic advances utilizing hierarchical data structures have resulted in a dramatic reduction in the time required for computer simulation of N-body systems with long-range interactions. Computations which required O($N\sp2$) operations can now be done in O(N log N) or O(N). We review these tree methods and find that they may be distinguished based on a few simple features. The Barnes-Hut (BH) algorithm has received a great deal of attention, and is the subject of the remainder of the dissertation. We present a generalization of the BH tree and analyze the statistical properties of such trees in detail. We also consider the expected number of operations entailed by an execution of the BH algorithm. We find an optimal value for m, the maximum number of bodies in a terminal cell, and confirm that the number of operations is O(N log N), even if the distribution of bodies is not uniform. The mathematical basis of all hierarchical methods is the multipole approximation. We discuss multipole approximations, for the case of arbitrary, spherically symmetric, and Newtonian Green's functions. We describe methods for computing multipoles and evaluating multipole approximations in each of these cases, emphasizing the tradeoff between generality and algorithmic complexity. N-body simulations in computational astrophysics can require 10$\sp6$ or even more bodies. Algorithmic advances are not sufficient, in and of themselves, to make computations of this size feasible. Parallel computation offers, a priori, the necessary computational power in terms of speed and memory. We show how the BH algorithm can be adapted to execute in parallel. We use orthogonal recursive bisection to partition space. The logical communication structure that emerges is that of a hypercube. A local version of the BH tree is constructed in each processor by iteratively exchanging data along each edge of the logical hypercube. We obtain speedups in excess of 380 on a 512 processor system for simulations of galaxy mergers with 180000 bodies. We analyze the performance of the parallel version of the algorithm and find that the overhead is due primarily to interprocessor synchronization delays and redundant computation. Communication is not a significant factor.

Journal ArticleDOI
TL;DR: The algorithm used to determine the equivalence of probabilistic automata can also solve the path equivalence problem for nondeterministic finite automata without $\lambda $-transitions and the equivalences problem for unambiguous finite automATA in polynomial time.
Abstract: Two probabilistic automata are equivalent if for any string x, the two automata accept x with equal probability. This paper presents an $O((n_1 + n_2 )^4 )$ algorithm for determining whether two probabilistic automata $U_1 $ and $U_2 $ are equivalent, where $n_1 $ and $n_2 $ are the number of states in $U_1 $ and $U_2 $, respectively.This improves the best previous result, which showed that the problem was in $coNP$. The existence of this algorithm implies that the covering and equivalence problems for uninitiated probabilistic automata are also polynomial-time solvable. The algorithm used to determine the equivalence of probabilistic automata can also solve the path equivalence problem for nondeterministic finite automata without $\lambda $-transitions and the equivalence problem for unambiguous finite automata in polynomial time.This paper studies the approximate equivalence (or $\delta $-equivalence) problem for probabilistic automata. An algorithm for the approximate equivalence problem for positive p...