scispace - formally typeset
Search or ask a question
Journal ArticleDOI

Parallel computation for well-endowed rings and space-bounded probabilistic machines

30 Jul 1984-Information & Computation (Academic Press Professional, Inc.)-Vol. 58, Iss: 1, pp 113-136
TL;DR: The method introduces a generalization of the ring of integers, called well-endowed rings, which possesses a very efficient parallel implementation of the basic (+,−,×) ring operations.
Abstract: It is shown that a probabilistic Turing acceptor or transducer running within space bound S can be simulated by a time S2 parallel machine and therefore by a space S2 deterministic machine. (Previous simulations ran in space S6.) In order to achieve these simulations, known algorithms are extended for the computation of determinants in small arithmetic parallel time to computations having small Boolean parallel time, and this development is applied to computing the completion of stochastic matrices. The method introduces a generalization of the ring of integers, called well-endowed rings. Such rings possess a very efficient parallel implementation of the basic (+,−,×) ring operations.

Content maybe subject to copyright    Report

Citations
More filters
Book ChapterDOI
02 Jan 1991
TL;DR: In this paper, the authors discuss parallel algorithms for shared-memory machines and discuss the theoretical foundations of parallel algorithms and parallel architectures, and present a theoretical analysis of the appropriate logical organization of a massively parallel computer.
Abstract: Publisher Summary This chapter discusses parallel algorithms for shared-memory machines. Parallel computation is rapidly becoming a dominant theme in all areas of computer science and its applications. It is estimated that, within a decade, virtually all developments in computer architecture, systems programming, computer applications and the design of algorithms will be taking place within the context of parallel computation. In preparation for this revolution, theoretical computer scientists have begun to develop a body of theory centered on parallel algorithms and parallel architectures. As there is no consensus yet on the appropriate logical organization of a massively parallel computer, and as the speed of parallel algorithms is constrained as much by limits on interprocessor communication as it is by purely computational issues, it is not surprising that a variety of abstract models of parallel computation have been pursued. Closest to the hardware level are the VLSI models, which focus on the technological limits of today's chips, in which gates and wires are packed into a small number of planar layers.

812 citations

Journal ArticleDOI
TL;DR: An attempt is made to identify important subclasses of NC and give interesting examples in each subclass, and a new problem complete for deterministic polynomial time is given, namely, finding the lexicographically first maximal clique in a graph.
Abstract: The class NC consists of problems solvable very fast (in time polynomial in log n ) in parallel with a feasible (polynomial) number of processors. Many natural problems in NC are known; in this paper an attempt is made to identify important subclasses of NC and give interesting examples in each subclass. The notion of NC 1 -reducibility is introduced and used throughout (problem R is NC 1 -reducible to problem S if R can be solved with uniform log-depth circuits using oracles for S ). Problems complete with respect to this reducibility are given for many of the subclasses of NC . A general technique, the “parallel greedy algorithm,” is identified and used to show that finding a minimum spanning forest of a graph is reducible to the graph accessibility problem and hence is in NC 2 (solvable by uniform Boolean circuits of depth O (log 2 n ) and polynomial size). The class LOGCFL is given a new characterization in terms of circuit families. The class DET of problems reducible to integer determinants is defined and many examples given. A new problem complete for deterministic polynomial time is given, namely, finding the lexicographically first maximal clique in a graph. This paper is a revised version of S. A. Cook, (1983, in “Proceedings 1983 Intl. Found. Comut. Sci. Conf.,” Lecture Notes in Computer Science Vol. 158, pp. 78–93, Springer-Verlag, Berlin/New York).

686 citations

Proceedings ArticleDOI
01 Jan 1987
TL;DR: A new algorithm for finding a maximum matching in a general graph that its only computationally non-trivial step is the inversion of a single integer matrix, the isolating lemma, and other applications to parallel computation and randomized reductions are shown.
Abstract: We present a new algorithm for finding a maximum matching in a general graph. The special feature of our algorithm is that its only computationally non-trivial step is the inversion of a single integer matrix. Since this step can be parallelized, we get a simple parallel (RNC 2) algorithm. At the heart of our algorithm lies a probabilistic lemma, the isolating lemma. We show other applications of this lemma to parallel computation and randomized reductions.

667 citations

Book ChapterDOI
David S. Johnson1
02 Jan 1991
TL;DR: This chapter discusses the concepts needed for defining the complexity classes, a set of problems of related resource-based complexity that can be solved by an abstract machine M using O(f(n) of resource R, where n is the size of the input.
Abstract: Publisher Summary This chapter discusses the concepts needed for defining the complexity classes. A complexity class is a set of problems of related resource-based complexity. A typical complexity class has a definition of the form—the set of problems that can be solved by an abstract machine M using O(f(n)) of resource R , where n is the size of the input. The simpler complexity classes are defined by various factors. The type of computational problem in which the most commonly used problems are decision problems. However, complexity classes can be defined based on function problems, counting problems, optimization problems, promise problems, etc. The most common model of computation is the deterministic Turing machine, but many complexity classes are based on nondeterministic Turing machines, etc.

618 citations

Book
06 Apr 1995
TL;DR: In providing an up-to-date survey of parallel computing research from 1994, Topics in Parallel Computing will prove invaluable to researchers and professionals with an interest in the super computers of the future.
Abstract: This volume provides an ideal introduction to key topics in parallel computing. With its cogent overview of the essentials of the subject as well as lists of P -complete- and open problems, extensive remarks corresponding to each problem, a thorough index, and extensive references, the book will prove invaluable to programmers stuck on problems that are particularly difficult to parallelize. In providing an up-to-date survey of parallel computing research from 1994, Topics in Parallel Computing will prove invaluable to researchers and professionals with an interest in the super computers of the future.

533 citations


Cites methods from "Parallel computation for well-endow..."

  • ...Remarks: For nth degree polynomials p, q ∈ Q[x], computing gcd(p, q) is in NC 2 via an NC1 reduction to Determinant (Cook and Sethi [69], Borodin, Cook, and Pippenger [42])....

    [...]

References
More filters
Book
01 Jan 1969
TL;DR: The theory of formal languages as a coherent theory is presented and its relationship to automata theory is made explicit, including the Turing machine and certain advanced topics in language theory.
Abstract: From the Preface (See Front Matter for full Preface) The study of formal languages constitutes an important subarea of computer science. This area sprang to life around 1956 when Noam Chomsky gave a mathematical model of a grammar in connection with his study of natural languages. Shortly afterwards, the concept of a grammar was found to be of great importance to the programmer when the syntax of the programming language ALGOL was defined by a context-free grammar. This development led naturally to syntax-directed compiling and the concept of a compiler compiler. Since then a considerable flurry of activity has taken place, the results of which have related formal languages and automata theory to such an extent that it is impossible to treat the areas separately. By now, no serious study of computer science would be complete without a knowledge of the techniques and results from language and automata theory. This book presents the theory of formal languages as a coherent theory and makes explicit its relationship to automata. The book begins with an explanation of the notion of a finite description of a language. The fundamental descriptive device--the grammar--is explained, as well as its three major subclasses--regular, context-free, and context-sensitive grammars. The context-free grammars are treated in detail, and such topics as normal forms, derivation trees, and ambiguity are covered. Four types of automata equivalent to the four types of grammars are described. These automata are the finite automaton, the pushdown automaton, the linear bounded automaton, and the Turing machine. The Turing machine is covered in detail, and unsolvability of the halting problem shown. The book concludes with certain advanced topics in language theory--closure properties, computational complexity, deterministic pushdown automata, LR(k) grammars, stack automata, and decidability.

1,595 citations

Journal ArticleDOI
TL;DR: The amount of storage needed to simulate a nondeterministic tape bounded Turingmachine on a deterministic Turing machine is investigated and a specific set is produced, namely the set of all codings of threadable mazes, such that, if there is any set which distinguishes nondeter microscopic complexity classes from deterministic tape complexity classes, then this is one such set.

1,414 citations

Journal ArticleDOI
TL;DR: Sign-digit representations limit carry-propagation to one position to the left during the operations of addition and subtraction in digital computers and arithmetic operations with signed-digit numbers: addition, subtraction, multiplication, division and roundoff are discussed.
Abstract: This paper describes a class of number representations which are called signed-digit representations. Signed-digit representations limit carry-propagation to one position to the left during the operations of addition and subtraction in digital computers. Carry-propagation chains are eliminated by the use of redundant representations for the operands. Redundancy in the number representation allows a method of fast addition and subtraction in which each sum (or difference) digit is the function only of the digits in two adjacent digital positions of the operands. The addition time for signed-digit numbers of any length is equal to the addition time for two digits. The paper discusses the properties of signed-digit representations and arithmetic operations with signed-digit numbers: addition, subtraction, multiplication, division and roundoff. A brief discussion of logical design problems for a signed-digit adder concludes the presentation.

1,232 citations


"Parallel computation for well-endow..." refers background in this paper

  • ...However, if one allows a redundant representation, then a constant depth implementation for addition is known (Avizienis, 1961)....

    [...]

  • ...This representation is due to Avizienis (1961)....

    [...]

Journal ArticleDOI
TL;DR: A recurstve construction is used to obtain a product circuit for solving the prefix problem and a Boolean clrcmt which has depth 2[Iog2n] + 2 and size bounded by 14n is obtained for n-bit binary addmon.
Abstract: The prefix problem is to compute all the products x t o x2 . . . . o xk for i ~ k .~ n, where o is an associative operation A recurstve construction IS used to obtain a product circuit for solving the prefix problem which has depth exactly [log:n] and size bounded by 4n An application yields fast, small Boolean ctrcmts to simulate fimte-state transducers. By simulating a sequentml adder, a Boolean clrcmt which has depth 2[Iog2n] + 2 and size bounded by 14n Is obtained for n-bit binary addmon The size can be decreased significantly by permitting the depth to increase by an addmve constant

1,159 citations