scispace - formally typeset
Proceedings ArticleDOI

Router Buffer Caching for Managing Shared Cache Blocks in Tiled Multi-Core Processors

TLDR
Wang et al. as mentioned in this paper proposed a congestion management technique in the LLC that equips the NoC router with small storage to keep a copy of heavily shared cache blocks, and also propose a prediction classifier in LLC controller.
Abstract
Multiple cores in a tiled multi-core processor are connected using a network-on-chip mechanism. All these cores share the last-level cache (LLC). For large-sized LLCs, generally, non-uniform cache architecture design is considered, where the LLC is split into multiple slices. Accessing highly shared cache blocks from an LLC slice by several cores simultaneously results in congestion at the LLC, which in turn increases the access latency. To deal with this issue, we propose a congestion management technique in the LLC that equips the NoC router with small storage to keep a copy of heavily shared cache blocks. To identify highly shared cache blocks, we also propose a prediction classifier in the LLC controller. We implement our technique in Sniper, an architectural simulator for multi-core systems, and evaluate its effectiveness by running a set of parallel benchmarks. Our experimental results show that the proposed technique is effective in reducing the LLC access time.

read more

Citations
More filters
Journal ArticleDOI

NCDE: In-Network Caching for Directory Entries to Expedite Data Access in Tiled-Chip Multiprocessors

TL;DR: In this paper , the authors explore the opportunity of mitigating problems associated with shared data access via in-network caching for directory entries (NCDE), which can utilize every input port's virtual channels to hold directory entries.
Journal ArticleDOI

NCDE: In-Network Caching for Directory Entries to Expedite Data Access in Tiled-Chip Multiprocessors

- 01 Jan 2023 - 
TL;DR: In this paper , the authors explore the opportunity of mitigating problems associated with shared data access via in-network caching for directory entries (NCDE), which can utilize every input port's virtual channels to hold directory entries.
References
More filters
Proceedings ArticleDOI

Network caching for Chip Multiprocessors

TL;DR: This paper develops three network caching designs to reduce L1 miss latencies and demonstrates that network caching architecture provides good scalability and also provides robust performance.
Book ChapterDOI

In-Network Caching for Chip Multiprocessors

TL;DR: In the proposed technique, shared data from read response packets that pass through the router are cached in its data store to reduce number of hops required to service future read requests and have the potential to reduce memory access latency.
Proceedings ArticleDOI

Network Victim Cache: Leveraging Network-on-Chip for Managing Shared Caches in Chip Multiprocessors

TL;DR: The network victim cache architecture removes the directory structure from shared L2 caches and stores directory information for the blocks recently cached by L1 caches in the network interface components decreasing on-chip directory memory overhead and improves the scalability.