scispace - formally typeset
Search or ask a question

Showing papers on "Time complexity published in 1997"


Journal ArticleDOI
Madhu Sudan1
TL;DR: To the best of the knowledge, this is the first efficient (i.e., polynomial time bounded) algorithm which provides error recovery capability beyond the error-correction bound of a code for any efficient code.

864 citations


Journal ArticleDOI
TL;DR: A new computational model is presented, which differs from a QC only in that the result of a measurement is the expectation value of the observable, rather than a random eigenvalue thereof, which can solve nondeterministic polynomial-time complete problems inPolynomial time.
Abstract: A quantum computer (QC) can operate in parallel on all its possible inputs at once, but the amount of information that can be extracted from the result is limited by the phenomenon of wave function collapse. We present a new computational model, which differs from a QC only in that the result of a measurement is the expectation value of the observable, rather than a random eigenvalue thereof. Such an expectation value QC can solve nondeterministic polynomial-time complete problems in polynomial time. This observation is significant precisely because the computational model can be realized, to a certain extent, by NMR spectroscopy on macroscopic ensembles of quantum spins, namely molecules in a test tube. This is made possible by identifying a manifold of statistical spin states, called pseudo-pure states, the mathematical description of which is isomorphic to that of an isolated spin system. The result is a novel NMR computer that can be programmed much like a QC, but in other respects more closely resembles a DNA computer. Most notably, when applied to intractable combinatorial problems, an NMR computer can use an amount of sample, rather than time, which grows exponentially with the size of the problem. Although NMR computers will be limited by current technology to exhaustive searches over only 15 to 20 bits, searches over as much as 50 bits are in principle possible, and more advanced algorithms could greatly extend the range of applicability of such machines.

787 citations


Journal ArticleDOI
TL;DR: SEQUITUR as mentioned in this paper is an algorithm that infers a hierarchical structure from a sequence of discrete symbols by replacing repeated phrases with a grammatical rule that generates the phrase, and continuing this process recursively.
Abstract: SEQUITUR is an algorithm that infers a hierarchical structure from a sequence of discrete symbols by replacing repeated phrases with a grammatical rule that generates the phrase, and continuing this process recursively. The result is a hierarchical representation of the original sequence, which offers insights into its lexical structure. The algorithm is driven by two constraints that reduce the size of the grammar, and produce structure as a by-product. SEQUITUR breaks new ground by operating incrementally. Moreover, the method's simple structure permits a proof that it operates in space and time that is linear in the size of the input. Our implementation can process 50,000 symbols per second and has been applied to an extensive range of real world sequences.

650 citations


Journal ArticleDOI
TL;DR: It is shown that some basic linear control design problems are NP-hard, implying that, unless P=NP, they cannot be solved by polynomial time algorithms.
Abstract: We show that some basic linear control design problems are NP-hard, implying that, unless P=NP, they cannot be solved by polynomial time algorithms. The problems that we consider include simultaneous stabilization by output feedback, stabilization by state or output feedback in the presence of bounds on the elements of the gain matrix, and decentralized control. These results are obtained by first showing that checking the existence of a stable matrix in an interval family of matrices is NP-hard.

411 citations


Proceedings ArticleDOI
01 Aug 1997
TL;DR: This paper gives the first methods to obtain seed sets that are provably small in size based on a variant of the contour tree (or topographic change tree), and develops a simple approximation algorithm giving a seed set of size at most twice the size of the minimum once the contours tree is known.
Abstract: For 2D or 3D meshes that represent the domain of continuous function to the reals, the contours|or isosurfaces|of a speci ed value are an important way to visualize the function To nd such contours, a seed set can be used for the starting points from which the traversal of the contours can begin This paper gives the rst methods to obtain seed sets that are provably small in size They are based on a variant of the contour tree (or topographic change tree) We give a new, simple algorithm to compute such a tree in regular and irregular meshes that requires O(n logn) time in 2D for meshes with n elements, and in O(n) time in higher dimensions The additional storage overhead is proportial to the maximum size of any contour (linear in the worst case, but typically less) Given the contour tree, a minimum size seed set can be computed in roughly quadratic time Since in practice this can be excessive, we develop a simple approximation algorithm giving a seed set of size at most twice the size of the minimum It requires O(n log n) time and linear storage once the contour tree is known We also give experimental results, showing the size of the seed sets for several data sets

363 citations


Journal ArticleDOI
TL;DR: A design methodology of practical solution algorithms for generally $\NP$-hard problems when restricted to partial k-trees (graphs with treewidth bounded by k) is presented, which accounts for dependency on the parameter k of the computational complexity of the resulting algorithms.
Abstract: In this paper, we consider a large class of vertex partitioning problems and apply to them the theory of algorithm design for problems restricted to partial k-trees. We carefully describe the details of algorithms and analyze their complexity in an attempt to make the algorithms feasible as solutions for practical applications. We give a precise characterization of vertex partitioning problems, which include domination, coloring and packing problems, and their variants. Several new graph parameters are introduced as generalizations of classical parameters. This characterization provides a basis for a taxonomy of a large class of problems, facilitating their common algorithmic treatment and allowing their uniform complexity classification. We present a design methodology of practical solution algorithms for generally $\NP$-hard problems when restricted to partial k-trees (graphs with treewidth bounded by k). This "practicality" accounts for dependency on the parameter k of the computational complexity of the resulting algorithms. By adapting the algorithm design methodology on partial k-trees to vertex partitioning problems, we obtain the first algorithms for these problems with reasonable time complexity as a function of treewidth. As an application of the methodology, we give the first polynomial-time algorithm on partial k-trees for computation of the Grundy number.

340 citations


Book ChapterDOI
24 Jun 1997
TL;DR: This paper presents an intuitive model for measuring the similarity between two time series that takes into account outliers, different scaling functions, and variable sampling rates, and shows the naturalness of this notion of similarity.
Abstract: Similarity of objects is one of the crucial concepts in several applications, including data mining For complex objects, similarity is nontrivial to define In this paper we present an intuitive model for measuring the similarity between two time series The model takes into account outliers, different scaling functions, and variable sampling rates Using methods from computational geometry, we show that this notion of similarity can be computed in polynomial time Using statistical approximation techniques, the algorithms can be speeded up considerably We give preliminary experimental results that show the naturalness of the notion

336 citations


Proceedings ArticleDOI
01 Jun 1997
TL;DR: The approach to answering range-max queries is based on precomputed max over balanced hierarchical tree structures, and a branch-and-bound-like procedure is used to speed up the finding of max in a region.
Abstract: A range query applies an aggregation operation over all selected cells of an OLAP data cube where the selection is specified by providing ranges of values for numeric dimensions. We present fast algorithms for range queries for two types of aggregation operations: SUM and MAX. These two operations cover techniques required for most popular aggregation operations, such as those supported by SQL.For range-sum queries, the essential idea is to precompute some auxiliary information (prefix sums) that is used to answer ad hoc queries at run-time. By maintaining auxiliary information which is of the same size as the data cube, all range queries for a given cube can be answered in constant time, irrespective of the size of the sub-cube circumscribed by a query. Alternatively, one can keep auxiliary information which is 1/bd of the size of the d-dimensional data cube. Response to a range query may now require access to some cells of the data cube in addition to the access to the auxiliary information, but the overall time complexity is typically reduced significantly. We also discuss how the precomputed information is incrementally updated by batching updates to the data cube. Finally, we present algorithms for choosing the subset of the data cube dimensions for which the auxiliary information is computed and the blocking factor to use for each such subset.Our approach to answering range-max queries is based on precomputed max over balanced hierarchical tree structures. We use a branch-and-bound-like procedure to speed up the finding of max in a region. We also show that with a branch-and-bound procedure, the average-case complexity is much smaller than the worst-case complexity.

335 citations


Proceedings ArticleDOI
S. Cameron1
20 Apr 1997
TL;DR: New results are presented that confirm an almost-constant time complexity for an enhanced version of Gilbert, Johnson and Keerthi's (GJK) algorithm, and also describe modifications to the algorithm to compute measures of penetration distance.
Abstract: The problem of tracking the distance between two convex polyhedra is finding applications in many areas of robotics, including intersection detection, collision detection, and path planning. We present new results that confirm an almost-constant time complexity for an enhanced version of Gilbert, Johnson and Keerthi's (GJK) algorithm, and also describe modifications to the algorithm to compute measures of penetration distance.

317 citations


Journal ArticleDOI
TL;DR: Time-efficient algorithms to solve the maze-routing problem on a reconfigurable mesh architecture and a fast algorithm to find the single shortest path (SSP) are presented.

275 citations


Journal ArticleDOI
TL;DR: The aim is to derive a class of objective functions the computation of which and the corresponding weight updates can be done in O(N) time, where N is the number of training patterns.
Abstract: In this paper, we study a number of objective functions for training new hidden units in constructive algorithms for multilayer feedforward networks. The aim is to derive a class of objective functions the computation of which and the corresponding weight updates can be done in O(N) time, where N is the number of training patterns. Moreover, even though input weight freezing is applied during the process for computational efficiency, the convergence property of the constructive algorithms using these objective functions is still preserved. We also propose a few computational tricks that can be used to improve the optimization of the objective functions under practical situations. Their relative performance in a set of two-dimensional regression problems is also discussed.

Journal ArticleDOI
TL;DR: The theoretical foundations of the binarization process studying the combinatorial optimization problems related to the minimization of the number of binary variables are developed and compact linear integer programming formulations of them are constructed.
Abstract: “Logical analysis of data” (LAD) is a methodology developed since the late eighties, aimed at discovering hidden structural information in data sets. LAD was originally developed for analyzing binary data by using the theory of partially defined Boolean functions. An extension of LAD for the analysis of numerical data sets is achieved through the process of “binarization” consisting in the replacement of each numerical variable by binary “indicator” variables, each showing whether the value of the original variable is above or below a certain level. Binarization was successfully applied to the analysis of a variety of real life data sets. This paper develops the theoretical foundations of the binarization process studying the combinatorial optimization problems related to the minimization of the number of binary variables. To provide an algorithmic framework for the practical solution of such problems, we construct compact linear integer programming formulations of them. We develop polynomial time algorithms for some of these minimization problems, and prove NP-hardness of others.

Journal ArticleDOI
TL;DR: In the course of solving the duplicate-grouping problem, a new universal class of hash functions of independent interest is described, and it is shown that both of the foregoing problems can be solved by randomized algorithms that useO(n) space and finish inO( n) time with probability tending to 1 asngrows to infinity.

Journal ArticleDOI
TL;DR: This work derives the first polynomial-time approximation scheme for this problem of assigning a set of n jobs to a system of m identical parallel machines so as to maximize the earliest machine completion time (without using idle times).

Journal ArticleDOI
TL;DR: In this paper, the minimum cost flow problem is solved in O(min(n 2m lognC, n 2m 2m2 logn) time, wheren is the number of nodes in the network, m is the maximum absolute arc costs, and C denotes the maximum arc costs if arc costs are integer and ∞ otherwise.
Abstract: Developing a polynomial time primal network simplex algorithm for the minimum cost flow problem has been a long standing open problem. In this paper, we develop one such algorithm that runs in O(min(n 2m lognC, n 2m2 logn)) time, wheren is the number of nodes in the network,m is the number of arcs, andC denotes the maximum absolute arc costs if arc costs are integer and ∞ otherwise. We first introduce a pseudopolynomial variant of the network simplex algorithm called the “premultiplier algorithm”. We then develop a cost-scaling version of the premultiplier algorithm that solves the minimum cost flow problem in O(min(nm lognC, nm 2 logn)) pivots. With certain simple data structures, the average time per pivot can be shown to be O(n). We also show that the diameter of the network polytope is O(nm logn).

Journal ArticleDOI
TL;DR: This paper describes a simple solution to this dilemma: limit the depth of partitioning, and for subproblems that exceed the limit switch to another algorithm with a better worst‐case bound.
Abstract: Quicksort is the preferred in-place sorting algorithm in many contexts, since its average computing time on uniformly distributed inputs is Θ(N log N), and it is in fact faster than most other sorting algorithms on most inputs. Its drawback is that its worst-case time bound is Θ(N2). Previous attempts to protect against the worst case by improving the way quicksort chooses pivot elements for partitioning have increased the average computing time too much – one might as well use heapsort, which has a Θ(N log N) worst-case time bound, but is on the average 2–5 times slower than quicksort. A similar dilemma exists with selection algorithms (for finding the i-th largest element) based on partitioning. This paper describes a simple solution to this dilemma: limit the depth of partitioning, and for subproblems that exceed the limit switch to another algorithm with a better worst-case bound. Using heapsort as the ‘stopper’ yields a sorting algorithm that is just as fast as quicksort in the average case, but also has an Θ(N log N) worst case time bound. For selection, a hybrid of Hoare's FIND algorithm, which is linear on average but quadratic in the worst case, and the Blum–Floyd–Pratt–Rivest–Tarjan algorithm is as fast as Hoare's algorithm in practice, yet has a linear worst-case time bound. Also discussed are issues of implementing the new algorithms as generic algorithms, and accurately measuring their performance in the framework of the C+:+ Standard Template Library. ©1997 by John Wiley & Sons, Ltd.

Journal ArticleDOI
TL;DR: In this article, a polynomial-time algorithm was proposed to find a proper 3-coloring of G3n,p,3 with high probability, whenever p ≥ c/n, where c is a sufficiently large absolute constant.
Abstract: Let G3n,p,3 be a random 3-colorable graph on a set of 3n vertices generated as follows. First, split the vertices arbitrarily into three equal color classes, and then choose every pair of vertices of distinct color classes, randomly and independently, to be edges with probability p. We describe a polynomial-time algorithm that finds a proper 3-coloring of G3n,p,3 with high probability, whenever p $\geq$ c/n, where c is a sufficiently large absolute constant. This settles a problem of Blum and Spencer, who asked if an algorithm can be designed that works almost surely for p $\geq$ polylog(n)/n [J. Algorithms, 19 (1995), pp. 204--234]. The algorithm can be extended to produce optimal k-colorings of random k-colorable graphs in a similar model as well as in various related models. Implementation results show that the algorithm performs very well in practice even for moderate values of c.

Journal ArticleDOI
TL;DR: This paper studies a two-machine flowshop scheduling problem with an availability constraint, and proves that the problem is NP-hard, and proposes two O(n log n) time heuristic algorithms to solve the problem optimally.

Journal ArticleDOI
TL;DR: Algorithms are given to construct a parameterized suffix tree in linear time and to find all maximal parameterized matches over a threshold length in a parameterization p-string in time linear in the size of the input plus the number of matches reported.
Abstract: As an aid in software maintenance, it would be useful to be able to track down duplication in large software systems efficiently. Duplication in code is often in the form of sections of code that are the same except for a systematic change of parameters such as identifiers and constants. To model such parameterized duplication in code, this paper introduces the notions of parameterized strings and parameterized matches of parameterized strings. A data structure called a parameterized suffix tree is defined to aid in searching for parameterized matches. For fixed alphabets, algorithms are given to construct a parameterized suffix tree in linear time and to find all maximal parameterized matches over a threshold length in a parameterized p-string in time linear in the size of the input plus the number of matches reported. The algorithms have been implemented, and experimental results show that they perform well on C code.

Journal ArticleDOI
TL;DR: The terminology of the most recent algorithm, Ukkonen's on-line construction, is used to explain its historic predecessors, revealing relationships much closer than one would expect, since the three algorithms are based on rather different intuitive ideas.
Abstract: We review the linear-time suffix tree constructions by Weiner, McCreight, and Ukkonen. We use the terminology of the most recent algorithm, Ukkonen's on-line construction, to explain its historic predecessors. This reveals relationships much closer than one would expect, since the three algorithms are based on rather different intuitive ideas. Moreover, it completely explains the differences between these algorithms in terms of simplicity, efficiency, and implementation complexity.

Journal ArticleDOI
TL;DR: Presents an optimal solution to the problem of allocating communicating periodic tasks to heterogeneous processing nodes (PNs) in a distributed real-time system subject to the precedence constraints resulting from intercommunication among the tasks to be allocated.
Abstract: Presents an optimal solution to the problem of allocating communicating periodic tasks to heterogeneous processing nodes (PNs) in a distributed real-time system. The solution is optimal in the sense of minimizing the maximum normalized task response time, called the system hazard, subject to the precedence constraints resulting from intercommunication among the tasks to be allocated. Minimization of the system hazard ensures that the solution algorithm allocates tasks so as to meet all task deadlines under an optimal schedule, whenever such an allocation exists. The task system is modeled with a task graph (TG), in which computation and communication modules, communication delays and intertask precedence constraints are clearly described. Tasks described by this TG are assigned to PNs by using a branch-and-bound (B&B) search algorithm. The algorithm traverses a search tree whose leaves correspond to potential solutions to the task allocation problem. We use a bounding method that prunes, in polynomial time, nonleaf vertices that cannot lead to an optimal solution, while ensuring that the search path leading to an optimal solution will never be pruned. For each generated leaf vertex, we compute the exact cost using the algorithm developed by Peng and Shin (1993). The lowest-cost leaf vertex (one with the least system hazard) represents an optimal task allocation. Computational experiences and examples are provided to demonstrate the concept, utility and power of the proposed approach.

Journal ArticleDOI
TL;DR: It is shown that for a large subset of single-unit RASs, the optimal DAP can be obtained in real-time with a computational cost which is a polynomial function of the system size.
Abstract: The development of efficient deadlock avoidance policies (DAPs) for sequential resource allocation systems (RASs) is a problem of increasing interest in the scientific community, largely because of its relevance to the design of large-scale flexibly automated manufacturing systems. Much of the work on this problem existing in the literature is focused on the so-called single-unit RAS model, which is the simplest model in the considered class of RASs. Furthermore, due to a well-established result stating that, even for single-unit RASs, the computation of the maximally permissive DAP is intractable (NP-hard), many researchers (including our group) have focused on obtaining good suboptimal policies which are computationally tractable (scalable) and provably correct. In the first part of the paper, it is shown, however, that for a large subset (in fact, a majority) of single-unit RASs, the optimal DAP can be obtained in real-time with a computational cost which is a polynomial function of the system size (i.e., the number of resource types and the distinct route stages of the processes running through the system). The implications of this result for the entire class of single-unit RASs are also explored. With a result on the design of optimal DAPs for single-unit RASs, the second part of the paper concentrates on the development of scalable and provably correct DAPs for the more general case of conjunctive RASs.

Proceedings ArticleDOI
04 May 1997
TL;DR: Greatly improved performance ratios of 4/3 for the 3-Set Cover problem and 6/5 for the Color Saving problem in graphs without independent sets of size 4 are obtained and shown to be the best possible with semi-local optimization.
Abstract: In this thesis a powerful new approximation technique called semi-local optimization is introduced. It provides very natural heuristics that are distinctly more effective than those based on local optimization. With an appropriate metric, semi-local optimization can still be viewed as a local optimization, but it has the advantage of making global changes to an approximate solution in polynomial time. Semi-local optimization generalizes recent heuristics of Halldorsson for the 3-Set Cover problem, the Color Saving problem, and the k-Set Cover problem. Greatly improved performance ratios of 4/3 for the 3-Set Cover problem and 6/5 for the Color Saving problem in graphs without independent sets of size 4 are obtained and shown to be the best possible with semi-local optimization. Also, based on the result for the 3-Set Cover problem and a restricted greedy phase for big sets, the performance ratio, for the k-Set Cover problem is correspondingly improved to ${\cal H}\sb{k}-1/2$. This result is also tight with the semi-local optimization technique. For larger values of k, further improvement better than ${\cal H}\sb{k}-1/2$ are also possible when the greedy selection of big sets is replaced by a local optimization selection. In the Color Saving problem, when larger independent sets exist, an improvement of the performance ratio to ${240}\over{193}$ for general graphs can be obtained.

Book ChapterDOI
11 Jul 1997
TL;DR: A randomized polynomial time approximation scheme for Euclidean TSP in R/sup 2/ is presented that is substantially more efficient than the earlier scheme (1996) and the scheme of Mitchell (1996).
Abstract: We present a randomized polynomial time approximation scheme for Euclidean TSP in R/sup 2/ that is substantially more efficient than our earlier scheme (1996) (and the scheme of Mitchell (1996)). For any fixed c>1 and any set of n nodes in the plane, the new scheme finds a (1+1/c)-approximation to the optimum traveling salesman tour in O(n(logn)/sup O(c)/) time. (Our earlier scheme ran in n/sup O(C)/ time.) For points in R/sup d/ the algorithm runs in O(n(logn)/sup (O(/spl radic/dc)/d-1)) time. This time is polynomial (actually nearly linear) for every fixed c, d. Designing such a polynomial-time algorithm was an open problem (our earlier algorithm (1996) ran in superpolynomial time for d/spl ges/3). The algorithm generalizes to the same set of Euclidean problems handled by the previous algorithm, including Steiner Tree, /spl kappa/-TSP, /spl kappa/-MST, etc, although for /spl kappa/-TSP and /spl kappa/-MST the running time gets multiplied by /spl kappa/. We also use our ideas to design nearly-linear time approximation schemes for Euclidean versions of problems that are known to be in P, such as Minimum Spanning Tree and Min Cost Perfect Matching. All our algorithms can be derandomized, though the running time then increases by O(n/sup d/) in R/sup d/. They also have simple parallel implementations (say, in NC/sup 2/).

Journal ArticleDOI
TL;DR: This paper proposes a novel GA-based algorithm with an objective to simultaneously meet the goals of high performance, scalability, and fast running time and outperforms both heuristics while taking considerably less running time.

Journal ArticleDOI
TL;DR: The Motzkin-Strauss QP is characterized, revealing interesting underlying discrete structures, and its properties are investigated, which are polynomial time verifiable.
Abstract: Given a graph G whose adjacency matrix is A, the Motzkin-Strauss formulation of the Maximum-Clique Problem is the quadratic program max{xTAx ∣ xTe = 1, x ≥ 0}. It is well known that the global optimum value of this QP is 1-1/ωG, where ωG is the clique number of G. Here, we characterize the following properties pertaining to the above QP: 1 first order optimality, 2 second order optimality, 3 local optimality, 4 strict local optimality. These characterizations reveal interesting underlying discrete structures, and are polynomial time verifiable. A parametrization of the Motzkin-Strauss QP is then introduced and its properties are investigated. Finally, an extension of the Motzkin-Strauss formulation is provided for the weighted clique number of a graph.

Journal ArticleDOI
TL;DR: The following three optimizations are discussed: (1) inheriting dual variables and partial solutions during partitioning, (2) sorting subproblems by lower cost bounds before solving, and (3) partitioning in an optimized order.
Abstract: We describe an implementation of an algorithm due to Murty for determining a ranked set of solutions to assignment problems. The intended use of the algorithm is in the context of multitarget tracking, where it has been shown that real-time multitarget tracking is feasible for some problems, but many other uses of the algorithm are also possible. The following three optimizations are discussed: (1) inheriting dual variables and partial solutions during partitioning, (2) sorting subproblems by lower cost bounds before solving, and (3) partitioning in an optimized order. When used to find the 100 best solutions to random 100/spl times/100 assignment problems, these optimizations produce a speedup of over a factor of 20, finding all 100 solutions in about 0.6 s. For a random cost matrix, the average time complexity for finding k solutions to random N/spl times/N problems appears to be nearly linear in both k and N, for sufficiently large k.

Journal ArticleDOI
TL;DR: It is proved that the maximum homeomorphic agreement subtree problem is $\cal{NP}$-complete for three trees with unbounded degrees and an approximation algorithm of time O(kn5) for choosing the species that are not in a maximum agreement subtrees of a set of k trees is shown.
Abstract: The maximum agreement subtree approach is one method of reconciling different evolutionary trees for the same set of species. An agreement subtree enables choosing a subset of the species for whom the restricted subtree is equivalent (under a suitable definition) in all given evolutionary trees. Recently, dynamic programming ideas were used to provide polynomial time algorithms for finding a maximum homeomorphic agreement subtree of two trees. Generalizing these methods to sets of more than two trees yields algorithms that are exponential in the number of trees. Unfortunately, it turns out that in reality one is usually presented with more than two trees, sometimes as many as thousands of trees. In this paper we prove that the maximum homeomorphic agreement subtree problem is $\cal{NP}$-complete for three trees with unbounded degrees. We then show an approximation algorithm of time O(kn5) for choosing the species that are not in a maximum agreement subtree of a set of k trees. Our approximation is guaranteed to provide a set that is no more than 4 times the optimum solution. While the set of evolutionary trees may be large in practice, the trees usually have very small degrees, typically no larger than three. We develop a new method for finding a maximum agreement subtree of k trees, of which one has degree bounded by d. This new method enables us to find a maximum agreement subtree in time O(knd + 1+ n2d).

Journal ArticleDOI
TL;DR: By adapting their teacher/learner model to grammatical inference it is proved that languages given by context-free grammars, simple deterministic grammar, linear grammARS and nondeterministic finite automata are not identifiable in the limit from polynomial time and data.
Abstract: When concerned about efficient grammatical inference two issues are relevant: the first one is to determine the quality of the result, and the second is to try to use polynomial time and space. A typical idea to deal with the first point is to say that an algorithm performs well if it infers {\it in\ the\ limit} the correct language. The second point has led to debate about how to define polynomial time: the main definitions of polynomial inference have been proposed by Pitt and Angluin. We return in this paper to a definition proposed by Gold that requires a characteristic set of strings to exist for each grammar, and this set to be polynomial in the size of the grammar or automaton that is to be learned, where the size of the sample is the sum of the lengths of all strings it includes. The learning algorithm must also infer correctly as soon as the characteristic set is included in the data. We first show that this definition corresponds to a notion of teachability as defined by Goldman and Mathias. By adapting their teacher/learner model to grammatical inference we prove that languages given by context-free grammars, simple deterministic grammars, linear grammars and nondeterministic finite automata are not identifiable in the limit from polynomial time and data.

Book ChapterDOI
22 Jun 1997
TL;DR: In this paper, weak and branching bisimulation for fully probabilistic transition systems is introduced. And the authors give an algorithm to decide weak (and branching) bisimulations with a time complexity cubic in the number of states of the system.
Abstract: Bisimulations that abstract from internal computation have proven to be useful for verification of compositionally defined transition systems. In the literature of probabilistic extensions of such transition systems, similar bisimulations are rare. In this paper, we introduce weak and branching bisimulation for fully probabilistic systems, transition systems where nondeterministic branching is replaced by probabilistic branching. In contrast to the nondeterministic case, both relations coincide. We give an algorithm to decide weak (and branching) bisimulation with a time complexity cubic in the number of states of the fully probabilistic system. This meets the worst case complexity for deciding branching bisimulation in the nondeterministic case. In addition, the relation is shown to be a congruence with respect to the operators of PLSCCS, a lazy synchronous probabilistic variant of CCS. We illustrate that due to these properties, weak bisimulation provides all the crucial ingredients for mechanised compositional veri�cation of probabilistic transition systems.