scispace - formally typeset
Proceedings ArticleDOI

1.1 Computing's energy problem (and what we can do about it)

Reads0
Chats0
TLDR
If the drive for performance and the end of voltage scaling have made power, and not the number of transistors, the principal factor limiting further improvements in computing performance, a new wave of innovative and efficient computing devices will be created.
Abstract
Our challenge is clear: The drive for performance and the end of voltage scaling have made power, and not the number of transistors, the principal factor limiting further improvements in computing performance. Continuing to scale compute performance will require the creation and effective use of new specialized compute engines, and will require the participation of application experts to be successful. If we play our cards right, and develop the tools that allow our customers to become part of the design process, we will create a new wave of innovative and efficient computing devices.

read more

Citations
More filters
Proceedings ArticleDOI

Memflow: memory-driven data scheduling with datapath co-design in accelerators for large-scale inference applications

TL;DR: MemFlow, memory-driven data scheduling with datapath co-design in accelerators, is proposed to improve computing performance and reduce higher-level memory accesses.

An Investigation into the Usefulness of the Smart Watch Interface for University Students

TL;DR: The results demonstrated that the smart watch can be a useful interface when used to retrieve small amounts of information as a complimentary device to a paired smart phone.
Posted Content

dCSR: A Memory-Efficient Sparse Matrix Representation for Parallel Neural Network Inference

TL;DR: Delta-Compressed Storage Row (dCSR) as mentioned in this paper is a storage format for sparse matrices that allows for both low overhead storage and fast inference on embedded systems with wide SIMD units.

RM-NTT: An RRAM-Based Compute-in-Memory Number Theoretic Transform Accelerator

TL;DR: In this paper , a resistive random access memory (RRAM)-based compute-in-memory (CIM) system is proposed to accelerate NTT and inverse NTT (INTT) operations.
Proceedings ArticleDOI

Shift-BNN: Highly-Efficient Probabilistic Bayesian Neural Network Training via Memory-Friendly Pattern Retrieving

TL;DR: Shift-BNN as mentioned in this paper eliminates all the off-chip data transfer by Gaussian Random Variables (GRVs) through the reversed shifting of Linear Feedback Shift Registers (LFSRs) without incurring any training accuracy loss.
References
More filters
Journal ArticleDOI

Design of ion-implanted MOSFET's with very small physical dimensions

TL;DR: This paper considers the design, fabrication, and characterization of very small Mosfet switching devices suitable for digital integrated circuits, using dimensions of the order of 1 /spl mu/.
Book

Low Power Digital CMOS Design

TL;DR: The Hierarchy of Limits of Power J.D. Stratakos, et al., and Low Power Programmable Computation coauthored with M.B. Srivastava, provide a review of the main approaches to Voltage Scaling Approaches.
Journal ArticleDOI

Towards energy-proportional datacenter memory with mobile DRAM

TL;DR: This work architects server memory systems using mobile DRAM devices, trading peak bandwidth for lower energy consumption per bit and more efficient idle modes, and demonstrates 3-5× lower memory power, better proportionality, and negligible performance penalties for data-center workloads.
Related Papers (5)