scispace - formally typeset
T

Tao Zhang

Researcher at Pennsylvania State University

Publications -  30
Citations -  2314

Tao Zhang is an academic researcher from Pennsylvania State University. The author has contributed to research in topics: Dram & Memory controller. The author has an hindex of 15, co-authored 30 publications receiving 1734 citations. Previous affiliations of Tao Zhang include Alibaba Group & Santa Clara University.

Papers
More filters
Journal ArticleDOI

PRIME: a novel processing-in-memory architecture for neural network computation in ReRAM-based main memory

TL;DR: This work proposes a novel PIM architecture, called PRIME, to accelerate NN applications in ReRAM based main memory, and distinguishes itself from prior work on NN acceleration, with significant performance improvement and energy saving.
Proceedings ArticleDOI

Overcoming the challenges of crossbar resistive memory architectures

TL;DR: The architecture improves the performance of a system using ReRAM-based main memory by about 44% over a conservative baseline and 14% over an aggressive baseline on average, and has less than 10% performance degradation compared to an ideal DRAM-only system.
Journal ArticleDOI

NVMain 2.0: A User-Friendly Memory Simulator to Model (Non-)Volatile Memory Systems

TL;DR: A flexible memory simulator - NVMain 2.0, is introduced to help the community for modeling not only commodity DRAMs but also emerging memory technologies, such as die-stacked DRAM caches, non-volatile memories including multi-level cells (MLC), and hybrid non-Volatile plus DRAM memory systems.
Journal ArticleDOI

Half-DRAM: a high-bandwidth and low-power DRAM architecture from the rethinking of fine-grained activation

TL;DR: A novel memory architecture Half-DRAM is proposed, in which the DRAM array is reorganized to enable only half of a row being activated, which can achieve both significant performance improvement and power reduction, with negligible design overhead.
Proceedings ArticleDOI

Sparse Tensor Core: Algorithm and Hardware Co-Design for Vector-wise Sparse Neural Networks on Modern GPUs

TL;DR: A novel pruning algorithm is devised to improve the workload balance and reduce the decoding overhead of the sparse neural networks and new instructions and micro-architecture optimization are proposed in Tensor Core to adapt to the structurally sparse Neural networks.