scispace - formally typeset
Proceedings ArticleDOI

Instruction fetch energy reduction using loop caches for embedded applications with small tight loops

Reads0
Chats0
TLDR
This paper proposes using a small instruction buffer, also called a loop cache, to save power in caches, which has no address tag store and knows precisely whether the next instruction request will hit in the loop cache well ahead of time.
Abstract
A fair amount of work has been done in recent years on reducing power consumption in caches by using a small instruction buffer placed between the execution pipe and a larger main cache. These techniques, however, often degrade the overall system performance. In this paper, we propose using a small instruction buffer, also called a loop cache, to save power. A loop cache has no address tag store. It consists of a direct-mapped data array and a loop cache controller. The loop cache controller knows precisely whether the next instruction request will hit in the loop cache, well ahead of time. As a result, there is no performance degradation.

read more

Citations
More filters
Proceedings ArticleDOI

Multi-bit Error Tolerant Caches Using Two-Dimensional Error Coding

TL;DR: Two-dimensional (2D) error coding in embedded memories is proposed, a scalable multi-bit error protection technique to improve memory reliability and yield and it is shown that 2D error coding can correct clustered errors up to 32times32 bits with significantly smaller performance, area, and power overheads than conventional techniques.
Book

Computer Architecture Techniques for Power-Efficiency

TL;DR: This book aims to document some of the most important architectural techniques that were invented, proposed, and applied to reduce both dynamic power and static power dissipation in processors and memory hierarchies by focusing on their common characteristics.
Journal ArticleDOI

System-level power-aware design techniques in real-time systems

TL;DR: An overview of power-aware techniques for the real-time system engineer as well as an up-to-date reference list for the researcher are provided.
Journal Article

Heap data allocation to scratch-pad memory in embedded systems

TL;DR: This thesis presents the first-ever compile-time method for allocating a portion of a program's dynamic data to scratch-pad memory that for the first time has no software-caching tags, requires no run-time per-access extra address translation, and is able to minimize the profile dependence issues which plague all similar allocation methods.
Proceedings ArticleDOI

Guaranteeing Hits to Improve the Efficiency of a Small Instruction Cache

TL;DR: The tagless hit instruction cache (TH-IC) is proposed, a technique for completely eliminating the performance penalty associated with filter caches, as well as a further reduction in energy consumption due to not having to access the tag array on cache hits.
References
More filters
Proceedings ArticleDOI

The filter cache: an energy efficient memory structure

TL;DR: Experimental results across a wide range of embedded applications show that the filter cache results in improved memory system energy efficiency, and this work proposes to trade performance for power consumption by filtering cache references through an unusually small L1 cache.
Proceedings ArticleDOI

Low-power digital design

TL;DR: In this article, the authors used an energy-delay metric to compare many of the proposed techniques and provided insight into some of the basic trade-offs in low-power design, including trade speed for power, do not waste power, and find a lower power problem.
Proceedings ArticleDOI

Cache design trade-offs for power and performance optimization: a case study

TL;DR: This paper examines performance and power trade-offs in cache designs and the effectiveness of energy reduction for several novel cache design techniques targeted for low power.
Patent

Data processing system having a cache and method therefor

TL;DR: In this paper, a look ahead feature for the valid bit array is provided, such that during a read of the cache, the valid bits for a next instruction is checked with the same index used to read the current instruction, so that the program can remain active as long as the program is in a loop which can be contained entirely within the cache.
Proceedings ArticleDOI

Low-cost branch folding for embedded applications with small tight loops

TL;DR: This paper proposes a hardware technique for folding branches when executing small loops, based on the detection and utilization of certain short backward branch instructions (sbb).
Related Papers (5)