scispace - formally typeset
Search or ask a question

Showing papers by "Massoud Pedram published in 1996"


Journal Article•DOI•
TL;DR: An in-depth survey of CAD methodologies and techniques for designing low power digital CMOS circuits and systems is presented and the many issues facing designers at architectural, logical, and physical levels of design abstraction are described.
Abstract: Low power has emerged as a principal theme in today's electronics industry. The need for low power has caused a major paradigm shift in which power dissipation is as important as performance and area. This article presents an in-depth survey of CAD methodologies and techniques for designing low power digital CMOS circuits and systems and describes the many issues facing designers at architectural, logical, and physical levels of design abstraction. It reviews some of the techniques and tools that have been proposed to overcome these difficulties and outlines the future challenges that must be met to design low power, high performance systems.

550 citations


Proceedings Article•DOI•
12 Aug 1996
TL;DR: Experimental results show that using four supply voltage levels on a number of standard benchmarks, an average energy saving of 53% can be obtained compared to using one fixed supply voltage level.
Abstract: We present a dynamic programming technique for solving the multiple supply voltage scheduling problem in both non-pipelined and functionally pipelined data-paths. The scheduling problem refers to the assignment of a supply voltage level to each operation in a data flow graph so as to minimize the average energy consumption for given computation time or throughput constraints or both. The energy model is accurate and accounts for the input pattern dependencies, re-convergent fanout induced dependencies, and the energy cost of level shifters. Experimental results show that using four supply voltage levels on a number of standard benchmarks, an average energy saving of 53% (with a computation time constraint of 1.5 times the critical path delay) can be obtained compared to using one fixed supply voltage level.

166 citations


Journal Article•DOI•
TL;DR: It is demonstrated that the average switching activity in the circuit can be calculated using either entropy or informational energy averages and the proposed switching activity estimation technique does not require simulation and is thus extremely fast, yet produces sufficiently accurate estimates.
Abstract: This paper considers the problem of estimating the power consumption at logic and register-transfer levels of design from an information theoretical point of view. In particular, it is demonstrated that the average switching activity in the circuit can be calculated using either entropy or informational energy averages. For control circuits and random logic, the output entropy (informational energy) per bit is calculated as a function of the input entropy (informational energy) per bit and an implementation dependent information scaling factor. For data-path circuits, the output entropy (informational energy) is calculated from the input entropy (informational energy) using a compositional technique which has linear complexity in terms of the circuit size. Finally, from these input and output values, the entropy (informational energy) per circuit line is calculated and used as an estimate for the average switching activity. The proposed switching activity estimation technique does not require simulation and is thus extremely fast, yet produces sufficiently accurate estimates.

114 citations


Journal Article•DOI•
TL;DR: Algorithms for disjunctive and nondisjunctive decomposition of Boolean functions and Boolean methods for identifying common subfunctions from multiple Boolean functions are presented and results are presented.
Abstract: This paper presents algorithms for disjunctive and nondisjunctive decomposition of Boolean functions and Boolean methods for identifying common subfunctions from multiple Boolean functions. Ordered binary decision diagrams are used to represent and manipulate Boolean functions so that the proposed methods can be implemented concisely. These techniques are applied to the synthesis of look-up table based field programmable gate arrays and results are presented.

97 citations


Proceedings Article•DOI•
20 Sep 1996
TL;DR: This work presents a technique to estimate the power consumption in a functionally pipelined data path and formulates the power optimization problem as a max cost multi commodity flow problem and solves it optimally.
Abstract: We investigate the problem of minimizing the total power consumption during the binding of operations to functional units in a scheduled data path with functional pipelining and conditional branching for data intensive applications. We first present a technique to estimate the power consumption in a functionally pipelined data path and then formulate the power optimization problem as a max cost multi commodity flow problem and solve it optimally. Our proposed method can augment most high level synthesis algorithms as a post processing step for reducing power after the optimizations for area or speed have been completed. An average power savings of 28% has been observed after we apply our method to pipelined designs that have been optimized using conventional techniques.

85 citations


Proceedings Article•DOI•
01 Nov 1996
TL;DR: The designer is provided with options to either improve the accuracy or the execution time when using power macro-modeling in the context of RTL simulation, and a regression estimator is described to reduce the error of the macro- modeling approach.
Abstract: In this paper, we propose a statistical power evaluation framework at the RT-level. We first discuss the power macro-modeling formulation, and then propose a simple random sampling technique to alleviate the the overhead of macro-modeling during RTL simulation. Next, we describe a regression estimator to reduce the error of the macro-modeling approach. Experimental results indicate that the execution time of the simple random sampling combined with power macro-modeling is 50 X lower than that of conventional macro-modeling while the percentage error of regression estimation combined with power macro-modeling is 16 X lower than that of conventional macro-modeling. Hence, we provide the designer with options to either improve the accuracy or the execution time when using power macro-modeling in the context of RTL simulation.

77 citations


Journal Article•DOI•
TL;DR: This work describes the structure and properties of edge-valued binary-decision diagrams (EVBDDs), and presents a general algorithm for performing a variety of binary operations.
Abstract: We present a new data structure called edge-valued binary-decision diagrams (EVBDD). An EVBDD is a directed acyclic graph, that provides a canonical and compact representation of functions that involve both Boolean and integer quantities. In general, EVBDDs provide a more versatile and powerful representation than ordinary binary decision diagrams. We first describe the structure and properties of EVBDDs, and present a general algorithm for performing a variety of binary operations. Next, we describe an important extension of EVBDDs, called Structural EVBDDs, and show how they can be used for hierarchical verification.

47 citations


Proceedings Article•DOI•
01 Jun 1996
TL;DR: POSE provides a unified framework for specifying and maintaining power relevant circuit information and means of estimating power consumption of a circuit using different load models and gives a set of options for making area-power trade-offs during logic optimization.
Abstract: Recent trends in the semiconductor industry have resulted in an increasing demand for low power circuits. POSE is a step in providing the EDA community and academia with an environment and tool suite for automatic synthesis and optimization of low power circuits. POSE provides a unified framework for specifying and maintaining power relevant circuit information and means of estimating power consumption of a circuit using different load models. POSE also gives a set of options for making area-power trade-offs during logic optimization.

42 citations


Proceedings Article•DOI•
01 Jun 1996
TL;DR: This work proposes an input vector compaction technique that preserves the statistical properties of the original sequence and shows that a compaction ratio of 100X is achieved with less than 2% average error in the power estimates.
Abstract: Accurate power estimation is essential for low power digital CMOS circuit design. Power dissipation is input pattern dependent. To obtain an accurate power estimate, a large input vector set must be used which leads to very long simulation time. One solution is to generate a compact vector set that is representative of the original input vector set and can be simulated in a reasonable time. We propose an input vector compaction technique that preserves the statistical properties of the original sequence. Experimental results show that a compaction ratio of 100X is achieved with less than 2% average error in the power estimates.

41 citations


Book Chapter•DOI•
01 Jan 1996
TL;DR: A canonical and compact data structure, called Edge Valued Binary Decision Diagrams (EVBDD), for representing and manipulating pseudo Boolean functions (PBF), and an extension of EVBDDs which associates both an additive and a multiplicative weight with the true edges of the function graph.
Abstract: We describe a canonical and compact data structure, called Edge Valued Binary Decision Diagrams (EVBDD), for representing and manipulating pseudo Boolean functions (PBF). EVBDDs are particularly useful when both arithmetic and Boolean operations are required. We describe a general algorithm on EVBDDs for performing any binary operation that is closed over the integers. Next, we discuss the relation between the probability expression of a Boolean function and its representation as a pseudo Boolean function. Utilizing this, we present algorithms for computing the probability spectrum and the Reed-Muller spectrum of a Boolean function directly on the EVBDD. Finally, we describe an extension of EVBDDs which associates both an additive and a multiplicative weight with the true edges of the function graph.

39 citations


Proceedings Article•DOI•
01 Jun 1996
TL;DR: Based on Moore-type machines, a general procedure for SSM synthesis is revealed and a new framework for sequence characterization is built to match designer's needs for sequence generation or compaction.
Abstract: The problem of stochastic sequential machines (SSM) synthesis is addressed and its relationship with the constrained sequence generation problem which arises during power estimation is discussed. In power estimation, one has to generate input vector sequences that satisfy a given statistical behavior (in terms of transition probabilities and correlations among bits) and/or to make these sequences as short as possible so as to improve the efficiency of power simulators, SSMs can be used to solve both problems. Based on Moore-type machines, a general procedure for SSM synthesis is revealed and a new framework for sequence characterization is built to match designer's needs for sequence generation or compaction. As results demonstrate, compaction ratios of 1-2 orders of magnitude can be obtained without much loss in accuracy of total power estimates.

Proceedings Article•DOI•
10 Nov 1996
TL;DR: The designer is provided with options to either improve the accuracy or the execution time when using power macro-modeling in the context of RTL simulation, and a regression estimator is described to reduce the error of the macro- modeling approach.
Abstract: In this paper, we propose a statistical power evaluation framework at the RT-level. We first discuss the power macro-modeling formulation, and then propose a simple random sampling technique to alleviate the the overhead of macro-modeling during RTL simulation. Next, we describe a regression estimator to reduce the error of the macro-modeling approach. Experimental results indicate that the execution time of the simple random sampling combined with power macro-modeling is 50 X lower than that of conventional macro-modeling while the percentage error of regression estimation combined with power macro-modeling is 16 X lower than that of conventional macro-modeling. Hence, we provide the designer with options to either improve the accuracy or the execution time when using power macro-modeling in the context of RTL simulation.

Journal Article•DOI•
TL;DR: It is shown that using don't cares computed for area optimization during local node minimization may result in an increase in the power consumption of other nodes in a Boolean network, and techniques for computing a subset of observability and satisfiability don't care conditions that can be used freely to optimize the local function of nodes are presented.
Abstract: This paper shows that using don't cares computed for area optimization during local node minimization may result in an increase in the power consumption of other nodes in a Boolean network. It then presents techniques for computing a subset of observability and satisfiability don't care conditions that can be used freely to optimize the local function of nodes. The concept of minimal variable support is then used to optimize the local function of each node for minimum power using its power relevant don't care set, that is, to reimplement the local function using a modified support that has a lower switching activity. Empirical results on a set of benchmark circuits are presented and discussed.

Proceedings Article•DOI•
11 Mar 1996
TL;DR: This paper presents an exact algorithm and two heuristics for solving the Bounded path length Minimal Spanning Tree (BMST) problem, which has polynomial space complexity and is hence more practical than the method presented by Gabow.
Abstract: This paper presents an exact algorithm and two heuristics for solving the Bounded path length Minimal Spanning Tree (BMST) problem. The exact algorithm which is based on iterative negative-sum-exchange(s) has polynomial space complexity and is hence more practical than the method presented by Gabow. The first heuristic method (BKRUS) is based on the classical Kruskal MST construction. For any given value of parameter /spl epsiv/, the algorithm constructs a routing tree with the longest interconnection path length at most (1+/spl epsiv/).R, and empirically with cost at most 1.19 times cost (BMST*) where R is the length of the direct path from the source to the farthest sink and BMST* is the optimal bounded path length MST. The second heuristic combines BKRUS and negative-sum-exchange(s) of depth 2 to improve results. Extensions of these techniques to the bounded path length Minimal Steiner Trees, using the Elmore delay model are presented as well. Empirical results demonstrate the effectiveness of these algorithms on a large benchmark set.

Proceedings Article•DOI•
01 Jun 1996
TL;DR: This paper presents a new approach for solving the Lower and Upper Bounded delay routing Tree (LUBT) problem using linear programming, and shows that the proposed method produces minimum cost LUBT for a given topology under a linear delay model.
Abstract: This paper presents a new approach for solving the Lower and Upper Bounded delay routing Tree (LUBT) problem using linear programming. LUBT is a Steiner tree rooted at the source node such that delays from the source to sink nodes lie between the given lower and upper bounds. We show that our proposed method produces minimum cost LUBT for a given topology under a linear delay model. Unlike recent works which control only the difference between the maximum and the minimum source-sink delay, we construct routing trees which satisfy distinct lower and upper bound constraints on the source-sink delays. This formulation exploits all the flexibility that is present in low power and high performance clock routing tree design.

Proceedings Article•DOI•
11 Mar 1996
TL;DR: This paper uses Boolean decomposition techniques to minimize the number of configurable logic blocks, the depth of the network and the power dissipations, and uses OBDDs to represent functions so that the methods can be implemented more effectively.
Abstract: In this paper, we address the problems of minimizing the area, delay and power during synthesis of field programmable gate arrays (FPGAs). We use Boolean decomposition techniques to minimize the number of configurable logic blocks (CLBs), the depth of the network and the power dissipations. We use OBDDs to represent functions so that our methods can be implemented more effectively. Our mapping algorithm is based on function decomposition which was pioneered by Ashenhurst [1959].

Book Chapter•DOI•
01 Jan 1996
TL;DR: Unless power consumption is dramatically reduced, the resulting heat will limit the feasible packing and performance of VLSI circuits and systems and circuits synthesized for low power are also less susceptible to run time failures.
Abstract: Low power, yet high-throughput and computationally intensive, circuits are becoming a critical application domain. One driving factor behind this trend is the growing class of personal computing devices (portable desktops, audio-and video-based multimedia products) and wireless communications systems (personal digital assistants and personal communicators) which demand high-speed computations and complex functionalities with low power consumption. Another crucial driving factor is that excessive power consumption is becoming the limiting factor in integrating more transistors on a single chip or on a multiple-chip module. Unless power consumption is dramatically reduced, the resulting heat will limit the feasible packing and performance of VLSI circuits and systems. Furthermore, circuits synthesized for low power are also less susceptible to run time failures.