scispace - formally typeset
Search or ask a question
Author

Alberto Sangiovanni-Vincentelli

Bio: Alberto Sangiovanni-Vincentelli is an academic researcher from University of California, Berkeley. The author has contributed to research in topics: Logic synthesis & Finite-state machine. The author has an hindex of 99, co-authored 934 publications receiving 45201 citations. Previous affiliations of Alberto Sangiovanni-Vincentelli include National University of Singapore & Lawrence Berkeley National Laboratory.


Papers
More filters
Journal ArticleDOI
01 Dec 1980
TL;DR: In this paper, the concept of nested macromodels is introduced to take advantage of the structural properties of large-scale integrated circuits, and a multilevel algorithm for the analysis of circuits contanining nested macro-models is presented and its implementation is discussed.
Abstract: The concept of nested macromodels is introduced to take advantage of the structural properties of large-scale integrated circuits. A multilevel algorithm for the analysis of circuits contanining nested macro-models is presented and its implementation is discussed.

3 citations

Book ChapterDOI
01 Jan 1993
TL;DR: This chapter describes a synthesis procedure that transforms a correct Signal Transition Graph specification with Complete State Coding into a logic circuit implementing it, and can be shown to be hazard-free, using the bounded wiredelay model.
Abstract: This chapter describes a synthesis procedure that transforms a correct Signal Transition Graph specification withComplete State Codinginto a logic circuit implementing it. The implementation can be shown to be hazard-free, using thebounded wiredelay model if: the circuit operates in an environment that obeys the STG specification, and the bounds on the delays are met by the circuit after manufacture.

3 citations

Proceedings ArticleDOI
30 Sep 2007
TL;DR: The panel will start with a brief overview on the state of the art in automotive networking followed by the panel statements and discussion.
Abstract: Summary form only given. For many years, the automotive network was a set of isolated field buses used for independent applications. With growing automotive electronics complexity, the network has become a means for systems integration. Now, gateways are used to allow the exchange of sensor data, diagnostics, and signalling for networked control. The approved design and certification process developed around the CAN bus does not easily scale to such larger networks and to interdependent embedded systems functions. With the new FlexRay bus standard, much higher bandwidth is provided. Combined with higher performance configurable gateways, there is a good chance that next generation network bandwidth requirements will be met. However, higher bandwidth does not necessarily mean better real-time or safety properties. Also, the automotive network is still widely seen as a collection of network components that are configured by the OEM to fit an individual set of automotive functions. How does this approach match the new automotive software standard, AUTOSAR, which defines a new level of interoperability and portability hiding much of the embedded platform properties from the application software? Would it be more appropriate to switch to an integrated automotive network that offers performance and safety guarantees based on formally defined performance and safety parameters and leave it to an independent network development how to reach such QoS data? Or are such integrated networks - as known from telecom - inappropriate and will be inferior, given the very complex function dependencies and cost pressure? There are many more questions in this context: are the current protocols, architectures, design methods, and tools appropriate? What innovations are most urgently needed? Who shall develop the networks in the future, the OEM or a 1st tier supplier? What would be the consequence for the design process? Do we need interoperable network service standards, e.g. as a complement to AUTOSAR? Will there be a unified automotive “internet protocol” that eventually dominates all communication in a car? How will future car-to-car communication be included in the automotive network strategy if it shall be used for real-time applications, such as in driver assistance systems? The panel will start with a brief overview on the state of the art in automotive networking followed by the panel statements and discussion.

3 citations

ReportDOI
21 Oct 2011
TL;DR: This paper addresses the problem of automatic synthesis, and in particular automatic and semantics-preserving implementation of Triggered Synchronous Block Diagrams (SBDs) on distributed, asynchronous execution platforms, and proposes methods to minimize this overhead.
Abstract: : Most of the design challenges for complex cyber-physical systems, where a digital controller governs a multiphysics plant, relate to the distributed nature of the systems to be controlled. Cars, airplanes, and power distribution grids are well-known examples. The characteristics of the communication network that connects the system components affect the derivation of the control law and the verification of design correctness. For this reason, there has been a strong interest in using languages and methodologies that facilitate the use of formal methods. These languages and methodologies are mostly based on a synchronous paradigm that, while satisfies the need for formalization, often results in an inefficient implementation requiring substantial overhead when compared to approaches that do not enforce synchronicity on the execution platform. Therefore, the interest is high for techniques that on one hand maintain the formal properties of synchronous models, and on the other hand, enable the use of asynchronous and distributed execution platforms with little overhead. In this paper we address the problem of automatic synthesis, and in particular automatic and semantics-preserving implementation of Triggered Synchronous Block Diagrams (SBDs) on distributed, asynchronous execution platforms. This problem was studied for pure SBDs (where all blocks are triggered in every synchronous step) in [23]. The method of [23] can be adapted to Triggered SBDs by using trigger elimination [16], where triggers are transformed to standard inputs. However, this often results in unnecessary communication overhead. In this paper we propose methods to minimize this overhead, thus improving the efficiency of the approach. We consider both general Triggered SBDs where the values of triggers are dynamically computed and are thus not known a-priori, as well as Timed SBDs where triggers are statically known, usually specified by (period, initial phase) pairs.

3 citations


Cited by
More filters
Journal ArticleDOI
01 Jan 1998
TL;DR: In this article, a graph transformer network (GTN) is proposed for handwritten character recognition, which can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters.
Abstract: Multilayer neural networks trained with the back-propagation algorithm constitute the best example of a successful gradient based learning technique. Given an appropriate network architecture, gradient-based learning algorithms can be used to synthesize a complex decision surface that can classify high-dimensional patterns, such as handwritten characters, with minimal preprocessing. This paper reviews various methods applied to handwritten character recognition and compares them on a standard handwritten digit recognition task. Convolutional neural networks, which are specifically designed to deal with the variability of 2D shapes, are shown to outperform all other techniques. Real-life document recognition systems are composed of multiple modules including field extraction, segmentation recognition, and language modeling. A new learning paradigm, called graph transformer networks (GTN), allows such multimodule systems to be trained globally using gradient-based methods so as to minimize an overall performance measure. Two systems for online handwriting recognition are described. Experiments demonstrate the advantage of global training, and the flexibility of graph transformer networks. A graph transformer network for reading a bank cheque is also described. It uses convolutional neural network character recognizers combined with global training techniques to provide record accuracy on business and personal cheques. It is deployed commercially and reads several million cheques per day.

42,067 citations

Journal ArticleDOI
Rainer Storn1, Kenneth Price
TL;DR: In this article, a new heuristic approach for minimizing possibly nonlinear and non-differentiable continuous space functions is presented, which requires few control variables, is robust, easy to use, and lends itself very well to parallel computation.
Abstract: A new heuristic approach for minimizing possibly nonlinear and non-differentiable continuous space functions is presented. By means of an extensive testbed it is demonstrated that the new method converges faster and with more certainty than many other acclaimed global optimization methods. The new method requires few control variables, is robust, easy to use, and lends itself very well to parallel computation.

24,053 citations

Journal ArticleDOI
01 Apr 1988-Nature
TL;DR: In this paper, a sedimentological core and petrographic characterisation of samples from eleven boreholes from the Lower Carboniferous of Bowland Basin (Northwest England) is presented.
Abstract: Deposits of clastic carbonate-dominated (calciclastic) sedimentary slope systems in the rock record have been identified mostly as linearly-consistent carbonate apron deposits, even though most ancient clastic carbonate slope deposits fit the submarine fan systems better. Calciclastic submarine fans are consequently rarely described and are poorly understood. Subsequently, very little is known especially in mud-dominated calciclastic submarine fan systems. Presented in this study are a sedimentological core and petrographic characterisation of samples from eleven boreholes from the Lower Carboniferous of Bowland Basin (Northwest England) that reveals a >250 m thick calciturbidite complex deposited in a calciclastic submarine fan setting. Seven facies are recognised from core and thin section characterisation and are grouped into three carbonate turbidite sequences. They include: 1) Calciturbidites, comprising mostly of highto low-density, wavy-laminated bioclast-rich facies; 2) low-density densite mudstones which are characterised by planar laminated and unlaminated muddominated facies; and 3) Calcidebrites which are muddy or hyper-concentrated debrisflow deposits occurring as poorly-sorted, chaotic, mud-supported floatstones. These

9,929 citations

Journal ArticleDOI
TL;DR: In this paper, the authors present a data structure for representing Boolean functions and an associated set of manipulation algorithms, which have time complexity proportional to the sizes of the graphs being operated on, and hence are quite efficient as long as the graphs do not grow too large.
Abstract: In this paper we present a new data structure for representing Boolean functions and an associated set of manipulation algorithms. Functions are represented by directed, acyclic graphs in a manner similar to the representations introduced by Lee [1] and Akers [2], but with further restrictions on the ordering of decision variables in the graph. Although a function requires, in the worst case, a graph of size exponential in the number of arguments, many of the functions encountered in typical applications have a more reasonable representation. Our algorithms have time complexity proportional to the sizes of the graphs being operated on, and hence are quite efficient as long as the graphs do not grow too large. We present experimental results from applying these algorithms to problems in logic design verification that demonstrate the practicality of our approach.

9,021 citations

Book
25 Apr 2008
TL;DR: Principles of Model Checking offers a comprehensive introduction to model checking that is not only a text suitable for classroom use but also a valuable reference for researchers and practitioners in the field.
Abstract: Our growing dependence on increasingly complex computer and software systems necessitates the development of formalisms, techniques, and tools for assessing functional properties of these systems. One such technique that has emerged in the last twenty years is model checking, which systematically (and automatically) checks whether a model of a given system satisfies a desired property such as deadlock freedom, invariants, and request-response properties. This automated technique for verification and debugging has developed into a mature and widely used approach with many applications. Principles of Model Checking offers a comprehensive introduction to model checking that is not only a text suitable for classroom use but also a valuable reference for researchers and practitioners in the field. The book begins with the basic principles for modeling concurrent and communicating systems, introduces different classes of properties (including safety and liveness), presents the notion of fairness, and provides automata-based algorithms for these properties. It introduces the temporal logics LTL and CTL, compares them, and covers algorithms for verifying these logics, discussing real-time systems as well as systems subject to random phenomena. Separate chapters treat such efficiency-improving techniques as abstraction and symbolic manipulation. The book includes an extensive set of examples (most of which run through several chapters) and a complete set of basic results accompanied by detailed proofs. Each chapter concludes with a summary, bibliographic notes, and an extensive list of exercises of both practical and theoretical nature.

4,905 citations